WO2022142108A1 - Method and apparatus for training interview entity recognition model, and method and apparatus for extracting interview information entity - Google Patents

Method and apparatus for training interview entity recognition model, and method and apparatus for extracting interview information entity Download PDF

Info

Publication number
WO2022142108A1
WO2022142108A1 PCT/CN2021/096583 CN2021096583W WO2022142108A1 WO 2022142108 A1 WO2022142108 A1 WO 2022142108A1 CN 2021096583 W CN2021096583 W CN 2021096583W WO 2022142108 A1 WO2022142108 A1 WO 2022142108A1
Authority
WO
WIPO (PCT)
Prior art keywords
interview
preset
label
recognition model
sample data
Prior art date
Application number
PCT/CN2021/096583
Other languages
French (fr)
Chinese (zh)
Inventor
邓悦
郑立颖
徐亮
Original Assignee
平安科技(深圳)有限公司
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 平安科技(深圳)有限公司 filed Critical 平安科技(深圳)有限公司
Publication of WO2022142108A1 publication Critical patent/WO2022142108A1/en

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F40/00Handling natural language data
    • G06F40/20Natural language analysis
    • G06F40/279Recognition of textual entities
    • G06F40/289Phrasal analysis, e.g. finite state techniques or chunking
    • G06F40/295Named entity recognition
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06QINFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES; SYSTEMS OR METHODS SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES, NOT OTHERWISE PROVIDED FOR
    • G06Q10/00Administration; Management
    • G06Q10/10Office automation; Time management
    • G06Q10/105Human resources

Definitions

  • the present application relates to the technical field of prediction models, and in particular, to a method, apparatus, device and medium for training an interview entity recognition model and extracting an interview information entity.
  • Named entity recognition is essentially a sequence labeling problem, that is, by inputting a sentence, outputting the entity corresponding to each word in the sentence, that is, identifying entities with specific meanings in the document, such as person names, place names, school names and proper nouns. For example, for the self-introduction of the interviewer in the intelligent recruitment process, it may be necessary to extract the company name and the name of the school, so as to facilitate the subsequent extraction and use of the interviewer's information.
  • the embodiments of the present application provide an interview entity recognition model training and interview information entity extraction method, device, equipment and medium, so as to solve the problem that supervised learning requires a large amount of data and unsupervised learning limits the accuracy of the model.
  • An interview entity recognition model training method comprising:
  • the preset interview sample data set includes at least one first interview sample data without an interview label
  • each auxiliary prediction module in the preset recognition model carry out auxiliary label prediction on the first interview sample data according to the interview coding vector, and obtain the auxiliary label distribution output by each of the auxiliary prediction modules;
  • a method for extracting interview information entities comprising:
  • the interview information of the target interviewee includes at least one interview sentence; one of the interview sentences includes a plurality of interview information words;
  • the interview sentence is input into the interview entity recognition model, and the interview information words in the interview sentence are extracted and recognized to obtain entity recognition results corresponding to the interview information words; the interview entity recognition model is Obtained according to the above interview entity recognition model training method;
  • An interview entity recognition model training device comprising:
  • a sample data acquisition module configured to acquire a preset interview sample data set;
  • the preset interview sample data set includes at least one first interview sample data without an interview label;
  • the standard label prediction module is used to input the first interview sample data into a preset recognition model including the first initial parameter, and perform the first interview sample data through the direct prediction module in the preset recognition model.
  • Standard label prediction to obtain standard label distribution and interview coding vector corresponding to the first interview sample data;
  • the auxiliary label prediction module is used to perform auxiliary label prediction on the first interview sample data according to the interview coding vector through each auxiliary prediction module in the preset recognition model, and obtain the output of each auxiliary prediction module.
  • a total loss value determination module configured to determine the total loss value of the preset recognition model according to each of the auxiliary label distributions and the standard label distribution;
  • a model training module configured to update and iterate the first initial parameter of the preset recognition model when the total loss value does not reach the preset convergence condition, until the total loss value reaches the preset convergence condition , and record the preset recognition model after convergence as an interview entity recognition model.
  • a device for extracting interview information entities comprising:
  • an interview information obtaining module used for obtaining interview information of a target interviewee; the interview information includes at least one interview sentence; one interview sentence includes a plurality of interview information words;
  • An entity extraction and recognition module is used to input the interview sentence into the interview entity recognition model, extract and recognize the interview information words in the interview sentence, and obtain entity recognition results corresponding to each of the interview information words;
  • the interview entity recognition model is obtained according to the above-mentioned interview entity recognition model training method;
  • An information storage module configured to insert the entity recognition result into a preset interview information storage template according to preset matching rules.
  • a computer device comprising a memory, a processor, and computer-readable instructions stored in the memory and executable on the processor, the processor implementing the following steps when executing the computer-readable instructions:
  • the preset interview sample data set includes at least one first interview sample data without an interview label
  • each auxiliary prediction module in the preset recognition model carry out auxiliary label prediction on the first interview sample data according to the interview coding vector, and obtain the auxiliary label distribution output by each of the auxiliary prediction modules;
  • a computer device comprising a memory, a processor, and computer-readable instructions stored in the memory and executable on the processor, the processor implementing the following steps when executing the computer-readable instructions:
  • the interview information of the target interviewee includes at least one interview sentence; one of the interview sentences includes a plurality of interview information words;
  • the interview sentence is input into the interview entity recognition model, and the interview information words in the interview sentence are extracted and recognized to obtain entity recognition results corresponding to the interview information words; the interview entity recognition model is Obtained according to the above interview entity recognition model training method;
  • One or more readable storage media storing computer-readable instructions that, when executed by one or more processors, cause the one or more processors to perform the following steps:
  • the preset interview sample data set includes at least one first interview sample data without an interview label
  • each auxiliary prediction module in the preset recognition model carry out auxiliary label prediction on the first interview sample data according to the interview coding vector, and obtain the auxiliary label distribution output by each of the auxiliary prediction modules;
  • One or more readable storage media storing computer-readable instructions that, when executed by one or more processors, cause the one or more processors to perform the following steps:
  • the interview information of the target interviewee includes at least one interview sentence; one of the interview sentences includes a plurality of interview information words;
  • the interview sentence is input into the interview entity recognition model, and the interview information words in the interview sentence are extracted and recognized to obtain entity recognition results corresponding to the interview information words; the interview entity recognition model is Obtained according to the above interview entity recognition model training method;
  • the above interview entity recognition model training, interview information entity extraction method, device, equipment and medium, the interview entity recognition model training method obtains a preset interview sample data set; the preset interview sample data set includes at least one that does not have an interview label.
  • the first interview sample data; the first interview sample data is input into a preset recognition model containing the first initial parameter, and the first interview sample data is analyzed by the direct prediction module in the preset recognition model.
  • Standard label prediction obtain standard label distribution and interview coding vector corresponding to the first interview sample data; through each auxiliary prediction module in the preset recognition model, according to the interview coding vector Perform auxiliary label prediction on the data to obtain auxiliary label distributions output by each of the auxiliary prediction modules; determine the total loss value of the preset recognition model according to each of the auxiliary label distributions and the standard label distribution; When the value does not reach the preset convergence condition, update and iterate the first initial parameter of the preset recognition model, until the total loss value reaches the preset convergence condition, the preset recognition model after convergence will be updated. Recorded as Interview Entity Recognition Model.
  • the auxiliary prediction module can give the model more different data features (such as each auxiliary prediction module).
  • the auxiliary label distribution output by the module improves the training efficiency of the interview entity recognition model and improves the accuracy of the trained interview entity recognition model.
  • FIG. 1 is a schematic diagram of an application environment of an interview entity recognition model training method and an interview information entity extraction method in an embodiment of the present application;
  • Fig. 2 is a flowchart of an interview entity recognition model training method in an embodiment of the present application
  • Fig. 3 is a flowchart of step S30 in the training method of interview entity recognition model in an embodiment of the present application
  • Fig. 4 is another flowchart of the training method of the interview entity recognition model in an embodiment of the present application.
  • FIG. 5 is a flowchart of a method for extracting interview information entities in an embodiment of the present application.
  • FIG. 6 is a schematic block diagram of an interview entity recognition model training device in an embodiment of the present application.
  • FIG. 7 is a schematic block diagram of an auxiliary label prediction module in an interview entity recognition model training device in an embodiment of the present application.
  • FIG. 8 is a schematic block diagram of an apparatus for extracting interview information entities in an embodiment of the present application.
  • FIG. 9 is a schematic diagram of a computer device in an embodiment of the present application.
  • the interview entity recognition model training method provided by the embodiment of the present application can be applied in the application environment shown in FIG. 1 .
  • the interview entity recognition model training method is applied in an interview entity recognition model training system.
  • the interview entity recognition model training system includes a client and a server as shown in FIG. 1 , and the client and the server communicate through a network for Solve the problem that supervised learning requires a large amount of data and unsupervised learning limits the accuracy of the model.
  • the client also known as the client, refers to the program corresponding to the server and providing local services for the client.
  • Clients can be installed on, but not limited to, various personal computers, laptops, smartphones, tablets, and portable wearable devices.
  • the server can be implemented as an independent server or a server cluster composed of multiple servers.
  • a method for training an interview entity recognition model is provided, and the method is applied to the server in FIG. 1 as an example for description, including the following steps:
  • S10 Obtain a preset interview sample data set; the preset interview sample data set includes at least one first interview sample data without an interview label;
  • the first interview sample data is data that does not have pre-marked interview labels; generally, a large amount of manually labeled data is required for model training and learning in supervised learning, but the demand for manually labeled data is large. , the manual labeling method wastes time and cannot output huge labeled data. Therefore, one of the problems to be solved in this application is how to train and learn the model more accurately and quickly in the absence of labeled data.
  • the first interview sample data can be selected according to different scenarios.
  • the first interview sample data can be the interviewee's self-introduction, or the interviewer's resume; In the editing scenario, the first interview sample data can be replaced with sentences in the movie script.
  • S20 Input the first interview sample data into a preset recognition model including the first initial parameter, and perform standard label prediction on the first interview sample data through the direct prediction module in the preset recognition model, to obtain Standard label distribution and interview coding vector corresponding to the first interview sample data;
  • the preset recognition model is a semi-supervised learning model formed by combining supervised learning and unsupervised learning;
  • the direct prediction module is obtained by training a small amount of data with interview labels, that is,
  • the direct prediction module is a module that has been trained.
  • the first interview sample data is used as the input of the direct prediction module, which includes a bidirectional recurrent neural network encoder.
  • the bidirectional cyclic neural network encoder is used for vector encoding the first interview sample data, and then obtains the interview encoding vector corresponding to the first interview sample data;
  • the direct prediction module also includes an annotation classification trained with interview annotation labels.
  • the labeling classifier is used to perform direct label prediction on the interview encoding vector to obtain the corresponding first interview sample data. Standard label distribution.
  • the auxiliary prediction module refers to a module that performs entity prediction on a certain word according to different word combinations.
  • the auxiliary prediction module is used to combine with the direct prediction module to form a semi-supervised mode. It should be noted that, in order to extract as much representation data of each word in the first interview sample data as possible, the features of the first interview sample data extracted by each auxiliary prediction module are all are different, that is, the basis for each auxiliary prediction module to discriminate the entities of the words in the first interview sample data is different, thereby improving the accuracy of the model's entity recognition; exemplarily, such as predicting words through the target Entity prediction is performed on the first three words, or entity prediction is performed on the last three words of the target prediction word, etc.
  • step S30 includes:
  • S301 Obtain at least two second forward coding vectors and at least two second reverse coding vectors in the interview coding vectors;
  • the second forward coding refers to performing the forward coding on each of the unlabeled sample words according to the normal
  • the second reverse encoding refers to encoding each of the unlabeled sample words in reverse order to obtain;
  • the unlabeled sample words refer to each segmented word obtained after segmenting the first interview sample data.
  • An auxiliary prediction module is associated with a second forward coding vector.
  • the encoder is a bidirectional cyclic neural network encoder, so the output encoding vector contains The second forward coding vector obtained by encoding the unlabeled sample words in the first interview sample data in the forward order (that is, starting from the first unlabeled sample word in the first interview sample data, from front to back coding), and the second reverse coding vector obtained by encoding each unlabeled sample word in the first interview sample data in reverse order (that is, starting from the last unlabeled sample word in the first interview sample data, Encoding from back to front).
  • obtaining at least two second forward coding vectors in the coding of the sample vector means that the second forward coding vector may be the first t unlabeled sample words in the unlabeled sample word.
  • a coding vector sequence composed of words; the second forward coding vector can also be a coding vector sequence composed of t-1 unlabeled sample words before the unlabeled sample word; similarly, at least two second reverse The coding vector refers to that the second reverse coding vector may be a coding vector sequence consisting of t unlabeled sample words after the unlabeled sample word; A sequence of encoded vectors consisting of t+1 unlabeled sample words.
  • S302 Determine the distribution of forward auxiliary labels corresponding to each of the unlabeled sample words according to each of the second forward coding vectors; Reverse auxiliary label distribution for words.
  • each unlabeled sample word in the corresponding second forward vector coding is Perform entity prediction to obtain the forward auxiliary label distribution corresponding to each second forward coding vector, that is, after a second forward coding vector is input to the auxiliary prediction module, a corresponding forward auxiliary label distribution will be output.
  • entity prediction is performed by each unlabeled sample word in the corresponding second reverse vector encoding, and the reverse auxiliary label distribution corresponding to each second reverse encoding vector is obtained.
  • the second forward encoding vector may be composed of t unlabeled sample words before the unlabeled sample word A. Therefore, in the auxiliary prediction module, entity prediction is performed on the unlabeled sample word A through the first t unlabeled sample words; The encoding vector sequence composed of t-1 unlabeled sample words, in another auxiliary prediction module, entity prediction is performed on the unlabeled sample word A through the first t-1 unlabeled sample words.
  • the distribution of the forward auxiliary labels output after the entity prediction of the unlabeled sample word A by the above-mentioned two second forward coding vectors is different, because the unlabeled sample words are compared to the unlabeled sample words through the first t unlabeled sample words.
  • the unlabeled sample word A can be touched; and when the unlabeled sample word A is predicted by the first t-1 unlabeled sample words, the unlabeled sample word A is also before the unlabeled sample word A.
  • auxiliary label prediction is performed on the first interview sample data according to the coding vector, and the auxiliary label distribution output by each auxiliary prediction module is obtained.
  • KL Kullback–Leibler divergence, relative entropy
  • q) refers to the KL divergence between the auxiliary label distribution and the interview label distribution;
  • p(x i ) represents the ith unlabeled sample word in the first interview sample data
  • q(x i ) represents the standard label distribution corresponding to the unlabeled sample words of p(x i ).
  • the total loss value of the preset recognition model is determined by the following expression:
  • L VCT ( ⁇ ) is the total loss value of the preset recognition model
  • is the number of the first interview sample data in the preset interview sample data set
  • k is the number of auxiliary prediction modules in the preset recognition model
  • x i ) is the standard label distribution corresponding to the i-th unlabeled sample word in the ⁇ -th first interview sample data
  • the convergence condition can be the condition that the total loss value is less than the set threshold, that is, when the total loss value is less than the set threshold, the training is stopped; the convergence condition can also be that the total loss value after 10,000 calculations is If the condition is very small and will not decrease again, that is, when the total loss value is small and will not decrease after 10,000 calculations, the training is stopped, and the preset recognition model after convergence is recorded as the interview entity recognition model.
  • the output results of the preset recognition model can be continuously approached to the accurate results, so that the recognition accuracy rate is getting higher and higher.
  • the preset recognition model after convergence is recorded as the interview entity recognition model.
  • the auxiliary prediction module can give the model more different data features (such as each The auxiliary label distribution output by the auxiliary prediction module) improves the training efficiency of the interview entity recognition model and improves the accuracy of the trained interview entity recognition model.
  • the interview entity recognition model may be stored in the blockchain.
  • Blockchain is a storage structure of encrypted and chained transactions formed by blocks.
  • the header of each block can include not only the hash values of all transactions in the block, but also the hash values of all transactions in the previous block, so that the transactions in the block can be tamper-proof based on the hash value.
  • anti-counterfeiting the newly generated transaction is filled into the block and after the consensus of the nodes in the blockchain network, it will be appended to the end of the blockchain to form a chain growth.
  • the preset interview sample data set also includes at least one second interview sample data with the interview label; Before the preset encoder in the recognition model performs standard label prediction on the first interview sample data, the method further includes:
  • S01 Input the second interview sample data into the preset recognition model, and perform direct label prediction on the second interview sample data through a preset prediction module that includes a second initial parameter in the preset recognition model , obtain the direct prediction label corresponding to the second interview sample data;
  • interview label refers to the manual labeling of each word in the second interview sample data in advance. For example, "I go to school in Peking University”, "Peking University” will be marked as a school in advance. Name entity.
  • the preset prediction module including the second parameter in the preset recognition model Perform direct label prediction on the interview sample data to obtain direct predicted labels corresponding to each word in the second interview sample data.
  • step S01 includes:
  • word segmentation processing may be performed on the second interview sample data through a stammering word segmentation method to obtain each marked sample word corresponding to the second interview sample data.
  • the encoder in the preset prediction module performs encoding processing on each of the labeled sample words to obtain a first forward encoding vector and a first reverse encoding vector;
  • the first forward encoding refers to The labeled sample words are obtained by encoding in a forward order;
  • the first reverse encoding refers to encoding each of the labeled sample words in a reverse order;
  • the encoder is a bidirectional cyclic neural network encoder, so the output encoding vector includes the first forward encoding vector obtained by encoding the labeled sample words in the second interview sample data in the forward order, and the second The first reverse encoding vector obtained by encoding each labeled sample word in the interview sample data in reverse order.
  • the labeling classifier in the preset prediction module is used to label each of the labelled sample words, and obtain the same label as each labelled sample.
  • the direct prediction label corresponding to the term.
  • the preset prediction module After the encoder in the preset prediction module performs encoding processing on each of the labeled sample words to obtain the first forward encoding vector and the first reverse encoding vector, the preset prediction module The labeling classifier performs label classification for each labeling sample word according to the first forward coding vector and the first reverse coding vector, and obtains a direct prediction label corresponding to each labeling sample word.
  • the second interview sample data is "I study at Peking University", where "Peking University” is manually marked as a school name entity, and then after the second interview sample data is input into the preset prediction module, the Each marked word (“I”, “Zai”, “Peking University”, “Learning”, etc.) in the second interview sample data is coded to obtain the first forward coding starting from “I” and coded from front to back The encoding vector, and the second reverse encoding vector encoded from the back to the front starting from “Peking University”, to classify "Peking University” according to the second forward encoding vector and the second reverse encoding vector.
  • the direct prediction label corresponding to "Peking University”.
  • S02 Determine the prediction loss value of the preset encoder according to the direct prediction label and the interview label
  • the labeling classifier in the preset prediction module performs label classification on each of the labeled sample words, and obtains the same After the direct prediction label corresponding to each labeled sample word, the direct prediction label corresponding to each labeled sample word is matched with the interview labeled label corresponding to each labeled sample word, and then the prediction loss value of the preset encoder is determined. .
  • the interview label includes a plurality of sample entity labels; in step S02, including:
  • the predicted loss value is determined by the cross entropy loss function.
  • the sample entity label is the label corresponding to each labeled sample word that exists in a second interview sample data, that is, there may be two or more labeled words in a second interview sample data. . Therefore, when the second interview sample data is input into the preset recognition model, the second interview sample data is directly labeled by the preset prediction module including the second initial parameter in the preset recognition model Predict, after obtaining the direct prediction label corresponding to the second interview sample data, obtain the sample entity label corresponding to each of the labeled sample words, according to the sample entity label corresponding to the same labeled sample word and the direct prediction label, to determine the label loss value corresponding to the labeled sample word; according to the label loss value corresponding to each labeled sample word, determine the predicted loss value through the cross entropy loss function.
  • predicted loss value can be determined according to the following expression:
  • L sup ( ⁇ ) is the prediction loss value
  • is the number of all the second interview sample data; refers to the sample entity label corresponding to the ⁇ th labeled sample word in the ⁇ th second interview sample data
  • a ⁇ ) refers to the ⁇ th labeled sample in the ⁇ th second interview sample data
  • CE() is the cross entropy loss function.
  • the training process in the training process, it first trains the direct prediction module through the second interview sample data with the interview label, and then uses the direct prediction module to analyze the first interview without the interview label through the direct prediction module.
  • the standard label prediction is performed on the interview sample data
  • the auxiliary label prediction is performed on the first interview sample data through each auxiliary prediction module, thereby characterizing the overall loss value of the interview entity recognition model in this application.
  • the convergence condition can be the condition that the predicted loss value is less than the set threshold, that is, when the predicted loss value is less than the set threshold, the training is stopped; the convergence condition can also be that the predicted loss value after 10,000 calculations is The condition that is very small and will not decrease again, that is, when the predicted loss value is small and will not decrease after 10,000 calculations, stop training, and record the preset prediction module after convergence as the direct prediction module .
  • the loss value adjusts the second initial parameter of the preset prediction module, and re-inputs the second interview sample data into the preset prediction module after adjusting the initial parameters, so that the predicted loss value corresponding to the second interview sample data reaches the preset value.
  • the convergence condition select another second interview sample data in the preset interview sample data set, and perform the above steps S01 to S03, and obtain the predicted loss value corresponding to the second interview sample data, and in the predicted loss value
  • the second initial parameter of the preset prediction module is adjusted again according to the predicted loss value, so that the predicted loss value corresponding to the second interview sample data reaches the preset convergence condition.
  • the preset prediction module after the preset prediction module is trained by using all the second interview sample data in the preset interview sample data set, the results output by the preset prediction module can be continuously approached to the accurate results, so that the recognition accuracy rate is getting higher and higher. Until all the prediction loss values corresponding to the second interview sample data reach the preset convergence condition, the preset prediction module after convergence is recorded as the direct prediction module.
  • the present application also proposes a method for extracting interview information entities, which is described by taking the method applied to the server in FIG. 1 as an example, including the following steps:
  • S60 Obtain interview information of the target interviewee; the interview information includes at least one interview sentence; one of the interview sentences includes a plurality of interview information words;
  • the interview information can be the information in the paper resume provided by the target interviewer, or it can be obtained by the target interviewee by taking the voice information of the target interviewer and converting it into text information during the self-introduction process.
  • S70 Input the interview sentence into the interview entity recognition model, extract and identify the interview information words in the interview sentence, and obtain entity recognition results corresponding to the interview information words; the interview entity recognition
  • the model is obtained according to the above-mentioned interview entity recognition model training method;
  • the entity recognition results include specific entity results, that is, the entity names corresponding to the interview information words are identified, such as school name entities, name entities, etc.; entity recognition results also include non-entity recognition results, such as " Interview information words such as "I” and "Yes” are non-entity.
  • the interview information words in the interview sentence are extracted and recognized, and the entity recognition result corresponding to each interview information word is obtained.
  • the recognition result is inserted into the preset interview information storage template according to the preset matching rules.
  • the preset interview information storage template includes multiple slots to be filled, for example, the template includes specific names, specific schools, specific genders, etc. to be filled, and then the entity results are matched with the slots to be filled, so as to "Peking University" is filled into the specific school slot, and then an interview information page corresponding to the target interviewee is formed.
  • an interview entity recognition model training device is provided, and the interview entity recognition model training device corresponds one-to-one with the interview entity recognition model training method in the above embodiment.
  • the interview entity recognition model training device includes a sample data acquisition module 10 , a standard label prediction module 20 , an auxiliary label prediction module 30 , a total loss value determination module 40 and a model training module 50 .
  • the detailed description of each functional module is as follows:
  • a sample data acquisition module 10 configured to acquire a preset interview sample data set; the preset interview sample data set includes at least one first interview sample data without an interview label;
  • the standard label prediction module 20 is used to input the first interview sample data into a preset recognition model including the first initial parameter, and the first interview sample data is analyzed by the direct prediction module in the preset recognition model. Perform standard label prediction to obtain standard label distribution and interview coding vector corresponding to the first interview sample data;
  • the auxiliary label prediction module 30 is configured to perform auxiliary label prediction on the first interview sample data according to the coding vector through each auxiliary prediction module in the preset recognition model, and obtain the output of each auxiliary prediction module.
  • a total loss value determination module 40 configured to determine the total loss value of the preset recognition model according to each of the auxiliary label distribution and the standard label distribution;
  • a model training module 50 configured to update and iterate the first initial parameter of the preset recognition model when the total loss value does not reach a preset convergence condition, until the total loss value reaches the preset convergence condition , the preset recognition model after convergence is recorded as the interview entity recognition model.
  • the interview entity recognition model training device further includes:
  • the direct label prediction module is used to input the second interview sample data into the preset recognition model, and the second interview sample is analyzed by the preset prediction module including the second initial parameter in the preset recognition model. performing direct label prediction on the data to obtain a direct prediction label corresponding to the second interview sample data;
  • a predicted loss value determination module configured to determine the predicted loss value of the preset encoder according to the direct prediction label and the interview label
  • a parameter update module configured to update and iterate the second initial parameter of the preset preset prediction module when the predicted loss value does not reach a preset convergence condition, until the predicted loss value reaches the preset convergence When conditions are met, the preset prediction module after convergence is recorded as the direct prediction module.
  • the direct label prediction module includes:
  • a word segmentation processing unit configured to perform word segmentation processing on the second interview sample data to obtain each labeled sample word corresponding to the second interview sample data;
  • an encoding processing unit configured to perform encoding processing on each of the labeled sample words by an encoder in the preset prediction module to obtain a first forward encoding vector and a first reverse encoding vector;
  • the first forward encoding Encoding refers to encoding each of the labeled sample words in a forward order;
  • the first reverse encoding refers to encoding each of the labeled sample words in a reverse order;
  • the label classification unit is configured to perform label classification on each of the labeled sample words by the labeling classifier in the preset prediction module according to the first forward coding vector and the first reverse coding vector, to obtain Directly predicted labels corresponding to each of the labeled sample words.
  • the predicted loss value determination module includes:
  • a sample entity label obtaining unit configured to obtain the sample entity label corresponding to each of the labeled sample words
  • a label loss value determination unit configured to determine a label loss value corresponding to the labeled sample word according to the sample entity label and the direct prediction label corresponding to the same labeled sample word;
  • the predicted loss value determining unit is configured to determine the predicted loss value through the cross entropy loss function according to the label loss value corresponding to each of the labeled sample words.
  • the auxiliary label prediction module 30 includes:
  • a coding vector obtaining unit 301 configured to obtain at least two second forward coding vectors and at least two second reverse coding vectors in the coding vectors;
  • the second forward coding refers to The words are obtained by encoding in a forward order;
  • the second reverse encoding refers to encoding each of the unlabeled sample words in a reverse order;
  • Auxiliary label distribution determining unit 302 configured to determine the distribution of forward auxiliary labels corresponding to each of the unlabeled sample words according to each of the second forward coding vectors; at the same time, determine according to the second reverse coding vector Reverse auxiliary label distribution corresponding to each of the unlabeled sample words.
  • a device for extracting interview information entities including:
  • the interview information obtaining module 60 is used to obtain the interview information of the target interviewee; the interview information includes at least one interview sentence; one of the interview sentences includes a plurality of interview information words;
  • the entity extraction and recognition module 70 is used to input the interview sentence into the interview entity recognition model, extract and recognize the interview information words in the interview sentence, and obtain the entity recognition result corresponding to each interview information word ; Described interview entity recognition model is obtained according to above-mentioned interview entity recognition model training method;
  • the information storage module 80 is configured to insert the entity recognition result into a preset interview information storage template according to preset matching rules.
  • Each module in the above-mentioned interview entity recognition model training device and interview information entity extraction device can be implemented in whole or in part by software, hardware and combinations thereof.
  • the above modules can be embedded in or independent of the processor in the computer device in the form of hardware, or stored in the memory in the computer device in the form of software, so that the processor can call and execute the operations corresponding to the above modules.
  • a computer device is provided, and the computer device may be a server, and its internal structure diagram may be as shown in FIG. 9 .
  • the computer device includes a processor, memory, a network interface, and a database connected by a system bus. Among them, the processor of the computer device is used to provide computing and control capabilities.
  • the memory of the computer device includes a readable storage medium, an internal memory.
  • the readable storage medium stores an operating system, computer readable instructions and a database.
  • the internal memory provides an environment for the execution of the operating system and computer-readable instructions in the readable storage medium.
  • the database of the computer device is used to store the data used in the interview entity recognition model training method in the above-mentioned embodiment, or the data used in the interview information entity extraction method in the above-mentioned embodiment.
  • the network interface of the computer device is used to communicate with an external terminal through a network connection.
  • the readable storage medium provided by this embodiment includes a non-volatile readable storage medium and a volatile readable storage medium.
  • a computer apparatus comprising a memory, a processor, and computer readable instructions stored in the memory and executable on the processor, the processor executing the computer readable instructions
  • the preset interview sample data set includes at least one first interview sample data without an interview label
  • each auxiliary prediction module in the preset recognition model carry out auxiliary label prediction on the first interview sample data according to the interview coding vector, and obtain the auxiliary label distribution output by each of the auxiliary prediction modules;
  • a computer apparatus comprising a memory, a processor, and computer readable instructions stored in the memory and executable on the processor, the processor executing the computer readable instructions
  • the interview information of the target interviewee includes at least one interview sentence; one of the interview sentences includes a plurality of interview information words;
  • the interview sentence is input into the interview entity recognition model, and the interview information words in the interview sentence are extracted and recognized to obtain entity recognition results corresponding to the interview information words; the interview entity recognition model is Obtained according to the above interview entity recognition model training method;
  • one or more readable storage media are provided storing computer-readable instructions that, when executed by one or more processors, cause the one or more processors to perform the following: step:
  • the preset interview sample data set includes at least one first interview sample data without an interview label
  • each auxiliary prediction module in the preset recognition model carry out auxiliary label prediction on the first interview sample data according to the interview coding vector, and obtain the auxiliary label distribution output by each of the auxiliary prediction modules;
  • one or more readable storage media are provided storing computer-readable instructions that, when executed by one or more processors, cause the one or more processors to perform the following: step:
  • the interview information of the target interviewee includes at least one interview sentence; one of the interview sentences includes a plurality of interview information words;
  • the interview sentence is input into the interview entity recognition model, and the interview information words in the interview sentence are extracted and recognized to obtain entity recognition results corresponding to the interview information words; the interview entity recognition model is Obtained according to the above interview entity recognition model training method;
  • Nonvolatile memory may include read only memory (ROM), programmable ROM (PROM), electrically programmable ROM (EPROM), electrically erasable programmable ROM (EEPROM), or flash memory.
  • Volatile memory may include random access memory (RAM) or external cache memory.
  • RAM is available in various forms such as static RAM (SRAM), dynamic RAM (DRAM), synchronous DRAM (SDRAM), double data rate SDRAM (DDRSDRAM), enhanced SDRAM (ESDRAM), synchronous chain Road (Synchlink) DRAM (SLDRAM), memory bus (Rambus) direct RAM (RDRAM), direct memory bus dynamic RAM (DRDRAM), and memory bus dynamic RAM (RDRAM), etc.

Landscapes

  • Engineering & Computer Science (AREA)
  • Business, Economics & Management (AREA)
  • Human Resources & Organizations (AREA)
  • Theoretical Computer Science (AREA)
  • Entrepreneurship & Innovation (AREA)
  • Strategic Management (AREA)
  • General Physics & Mathematics (AREA)
  • Physics & Mathematics (AREA)
  • Tourism & Hospitality (AREA)
  • Health & Medical Sciences (AREA)
  • Operations Research (AREA)
  • Marketing (AREA)
  • General Business, Economics & Management (AREA)
  • Economics (AREA)
  • Data Mining & Analysis (AREA)
  • Quality & Reliability (AREA)
  • Artificial Intelligence (AREA)
  • Audiology, Speech & Language Pathology (AREA)
  • Computational Linguistics (AREA)
  • General Health & Medical Sciences (AREA)
  • General Engineering & Computer Science (AREA)
  • Information Retrieval, Db Structures And Fs Structures Therefor (AREA)

Abstract

A method and apparatus for training an interview entity recognition model, and a method and apparatus for extracting an interview information entity. The method for training an interview entity recognition model comprises: performing standard label prediction on first interview sample data by means of a direct prediction module in a preset recognition model, so as to obtain a standard label distribution and an interview coding vector; performing auxiliary label prediction on the first interview sample data by means of each auxiliary prediction module and according to the coding vector, so as to obtain an auxiliary label distribution output by each auxiliary prediction module; determining a total loss value of the preset recognition model according to each auxiliary label distribution and the standard label distribution; and when the total loss value does not meet a preset convergence condition, updating and iterating a first initial parameter of the preset recognition model until the total loss value meets the preset convergence condition, and recording the preset recognition model after convergence is realized as an interview entity recognition model. By means of the method, the model training efficiency and the model recognition accuracy are improved.

Description

面试实体识别模型训练、面试信息实体提取方法及装置Interview entity recognition model training, interview information entity extraction method and device
本申请要求于2020年12月30日提交中国专利局、申请号为202011620124.1,发明名称为“面试实体识别模型训练、面试信息实体提取方法及装置”的中国专利申请的优先权,其全部内容通过引用结合在本申请中。This application claims the priority of the Chinese patent application filed on December 30, 2020 with the application number 202011620124.1 and the invention titled "Interview Entity Recognition Model Training, Interview Information Entity Extraction Method and Device", the entire contents of which are approved by Reference is incorporated in this application.
技术领域technical field
本申请涉及预测模型技术领域,尤其涉及一种面试实体识别模型训练、面试信息实体提取方法、装置、设备及介质。The present application relates to the technical field of prediction models, and in particular, to a method, apparatus, device and medium for training an interview entity recognition model and extracting an interview information entity.
背景技术Background technique
命名实体识别本质是一种序列标注问题,就是通过输入一句话,输出句子中的每个单词对应的实体,即识别出文档中具有特定意义的实体,例如人名,地名,学校名和专有名词。比如,在智能招聘流程中对于面试者的自我介绍,可能需要将公司名称,学校名称提取出来,进而方便后续对面试者的信息进行提取和使用。Named entity recognition is essentially a sequence labeling problem, that is, by inputting a sentence, outputting the entity corresponding to each word in the sentence, that is, identifying entities with specific meanings in the document, such as person names, place names, school names and proper nouns. For example, for the self-introduction of the interviewer in the intelligent recruitment process, it may be necessary to extract the company name and the name of the school, so as to facilitate the subsequent extraction and use of the interviewer's information.
发明人意识到,针对命名实体识别的问题,如果通过监督学习进行命名实体识别,则对于数据量的需求非常大,而通过无监督学习的算法,比如使用预训练语言模型,进行命名实体识别,则限制了模型的准确率。The inventor realized that for the problem of named entity recognition, if named entity recognition is performed through supervised learning, the demand for the amount of data is very large, and through unsupervised learning algorithms, such as using pretrained language models, named entity recognition, This limits the accuracy of the model.
申请内容Application content
本申请实施例提供一种面试实体识别模型训练、面试信息实体提取方法、装置、设备及介质,以解决监督学习对数据量的需求非常大,以及无监督学习限制模型准确率的问题。The embodiments of the present application provide an interview entity recognition model training and interview information entity extraction method, device, equipment and medium, so as to solve the problem that supervised learning requires a large amount of data and unsupervised learning limits the accuracy of the model.
一种面试实体识别模型训练方法,包括:An interview entity recognition model training method, comprising:
获取预设面试样本数据集;所述预设面试样本数据集中包含至少一个不具有面试标注标签的第一面试样本数据;obtaining a preset interview sample data set; the preset interview sample data set includes at least one first interview sample data without an interview label;
将所述第一面试样本数据输入至包含第一初始参数的预设识别模型中,通过所述预设识别模型中的直接预测模块对所述第一面试样本数据进行标准标签预测,得到标准标签分布以及与所述第一面试样本数据对应的面试编码向量;Inputting the first interview sample data into a preset recognition model including the first initial parameter, and performing standard label prediction on the first interview sample data through the direct prediction module in the preset recognition model to obtain a standard label distribution and an interview coding vector corresponding to the first interview sample data;
通过所述预设识别模型中的各辅助预测模块,根据所述面试编码向量对所述第一面试样本数据进行辅助标签预测,得到与各所述辅助预测模块输出的辅助标签分布;Through each auxiliary prediction module in the preset recognition model, carry out auxiliary label prediction on the first interview sample data according to the interview coding vector, and obtain the auxiliary label distribution output by each of the auxiliary prediction modules;
根据各所述辅助标签分布与所述标准标签分布确定所述预设识别模型的总损失值;Determine the total loss value of the preset recognition model according to each of the auxiliary label distribution and the standard label distribution;
在所述总损失值未达到预设的收敛条件时,更新迭代所述预设识别模型的第一初始参数,直至所述总损失值达到所述预设的收敛条件时,将收敛之后的所述预设识别模型记录为面试实体识别模型。When the total loss value does not reach the preset convergence condition, update and iterate the first initial parameter of the preset recognition model, until the total loss value reaches the preset convergence condition, all the The aforementioned preset recognition model is recorded as the interview entity recognition model.
一种面试信息实体提取方法,包括:A method for extracting interview information entities, comprising:
获取目标面试者的面试信息;所述面试信息中包含至少一个面试句子;一个所述面试句子中包含多个面试信息字词;Obtain the interview information of the target interviewee; the interview information includes at least one interview sentence; one of the interview sentences includes a plurality of interview information words;
将所述面试句子输入至面试实体识别模型中,对所述面试句子中的面试信息字词进行提取识别,得到与各所述面试信息字词对应的实体识别结果;所述面试实体识别模型是根据上述面试实体识别模型训练方法得到的;The interview sentence is input into the interview entity recognition model, and the interview information words in the interview sentence are extracted and recognized to obtain entity recognition results corresponding to the interview information words; the interview entity recognition model is Obtained according to the above interview entity recognition model training method;
将所述实体识别结果按照预设匹配规则插入预设面试信息存储模板。Insert the entity recognition result into a preset interview information storage template according to preset matching rules.
一种面试实体识别模型训练装置,包括:An interview entity recognition model training device, comprising:
样本数据获取模块,用于获取预设面试样本数据集;所述预设面试样本数据集中包含至少一个不具有面试标注标签的第一面试样本数据;a sample data acquisition module, configured to acquire a preset interview sample data set; the preset interview sample data set includes at least one first interview sample data without an interview label;
标准标签预测模块,用于将所述第一面试样本数据输入至包含第一初始参数的预设识别模型中,通过所述预设识别模型中的直接预测模块对所述第一面试样本数据进行标准标签预测,得到标准标签分布以及与所述第一面试样本数据对应的面试编码向量;The standard label prediction module is used to input the first interview sample data into a preset recognition model including the first initial parameter, and perform the first interview sample data through the direct prediction module in the preset recognition model. Standard label prediction, to obtain standard label distribution and interview coding vector corresponding to the first interview sample data;
辅助标签预测模块,用于通过所述预设识别模型中的各辅助预测模块,根据所述面试编码向量对所述第一面试样本数据进行辅助标签预测,得到与各所述辅助预测模块输出的辅助标签分布;The auxiliary label prediction module is used to perform auxiliary label prediction on the first interview sample data according to the interview coding vector through each auxiliary prediction module in the preset recognition model, and obtain the output of each auxiliary prediction module. Auxiliary label distribution;
总损失值确定模块,用于根据各所述辅助标签分布与所述标准标签分布确定所述预设识别模型的总损失值;a total loss value determination module, configured to determine the total loss value of the preset recognition model according to each of the auxiliary label distributions and the standard label distribution;
模型训练模块,用于在所述总损失值未达到预设的收敛条件时,更新迭代所述预设识别模型的第一初始参数,直至所述总损失值达到所述预设的收敛条件时,将收敛之后的所述预设识别模型记录为面试实体识别模型。A model training module, configured to update and iterate the first initial parameter of the preset recognition model when the total loss value does not reach the preset convergence condition, until the total loss value reaches the preset convergence condition , and record the preset recognition model after convergence as an interview entity recognition model.
一种面试信息实体提取装置,包括:A device for extracting interview information entities, comprising:
面试信息获取模块,用于获取目标面试者的面试信息;所述面试信息中包含至少一个面试句子;一个所述面试句子中包含多个面试信息字词;an interview information obtaining module, used for obtaining interview information of a target interviewee; the interview information includes at least one interview sentence; one interview sentence includes a plurality of interview information words;
实体提取识别模块,用于将所述面试句子输入至面试实体识别模型中,对所述面试句子中的面试信息字词进行提取识别,得到与各所述面试信息字词对应的实体识别结果;所述面试实体识别模型是根据上述面试实体识别模型训练方法得到的;An entity extraction and recognition module is used to input the interview sentence into the interview entity recognition model, extract and recognize the interview information words in the interview sentence, and obtain entity recognition results corresponding to each of the interview information words; The interview entity recognition model is obtained according to the above-mentioned interview entity recognition model training method;
信息存储模块,用于将所述实体识别结果按照预设匹配规则插入预设面试信息存储模板。An information storage module, configured to insert the entity recognition result into a preset interview information storage template according to preset matching rules.
一种计算机设备,包括存储器、处理器以及存储在所述存储器中并可在所述处理器上运行的计算机可读指令,所述处理器执行所述计算机可读指令时实现如下步骤:A computer device, comprising a memory, a processor, and computer-readable instructions stored in the memory and executable on the processor, the processor implementing the following steps when executing the computer-readable instructions:
获取预设面试样本数据集;所述预设面试样本数据集中包含至少一个不具有面试标注标签的第一面试样本数据;obtaining a preset interview sample data set; the preset interview sample data set includes at least one first interview sample data without an interview label;
将所述第一面试样本数据输入至包含第一初始参数的预设识别模型中,通过所述预设识别模型中的直接预测模块对所述第一面试样本数据进行标准标签预测,得到标准标签分布以及与所述第一面试样本数据对应的面试编码向量;Inputting the first interview sample data into a preset recognition model including the first initial parameter, and performing standard label prediction on the first interview sample data through the direct prediction module in the preset recognition model to obtain a standard label distribution and an interview coding vector corresponding to the first interview sample data;
通过所述预设识别模型中的各辅助预测模块,根据所述面试编码向量对所述第一面试样本数据进行辅助标签预测,得到与各所述辅助预测模块输出的辅助标签分布;Through each auxiliary prediction module in the preset recognition model, carry out auxiliary label prediction on the first interview sample data according to the interview coding vector, and obtain the auxiliary label distribution output by each of the auxiliary prediction modules;
根据各所述辅助标签分布与所述标准标签分布确定所述预设识别模型的总损失值;在所述总损失值未达到预设的收敛条件时,更新迭代所述预设识别模型的第一初始参数,直至所述总损失值达到所述预设的收敛条件时,将收敛之后的所述预设识别模型记录为面试实体识别模型。Determine the total loss value of the preset recognition model according to the distribution of each auxiliary label and the standard label distribution; when the total loss value does not reach the preset convergence condition, update the first step of iterating the preset recognition model. An initial parameter, until the total loss value reaches the preset convergence condition, the preset recognition model after convergence is recorded as the interview entity recognition model.
一种计算机设备,包括存储器、处理器以及存储在所述存储器中并可在所述处理器上运行的计算机可读指令,所述处理器执行所述计算机可读指令时实现如下步骤:A computer device, comprising a memory, a processor, and computer-readable instructions stored in the memory and executable on the processor, the processor implementing the following steps when executing the computer-readable instructions:
获取目标面试者的面试信息;所述面试信息中包含至少一个面试句子;一个所述面试句子中包含多个面试信息字词;Obtain the interview information of the target interviewee; the interview information includes at least one interview sentence; one of the interview sentences includes a plurality of interview information words;
将所述面试句子输入至面试实体识别模型中,对所述面试句子中的面试信息字词进行提取识别,得到与各所述面试信息字词对应的实体识别结果;所述面试实体识别模型是根据上述面试实体识别模型训练方法得到的;The interview sentence is input into the interview entity recognition model, and the interview information words in the interview sentence are extracted and recognized to obtain entity recognition results corresponding to the interview information words; the interview entity recognition model is Obtained according to the above interview entity recognition model training method;
将所述实体识别结果按照预设匹配规则插入预设面试信息存储模板。Insert the entity recognition result into a preset interview information storage template according to preset matching rules.
一个或多个存储有计算机可读指令的可读存储介质,所述计算机可读指令被一个或多个处理器执行时,使得所述一个或多个处理器执行如下步骤:One or more readable storage media storing computer-readable instructions that, when executed by one or more processors, cause the one or more processors to perform the following steps:
获取预设面试样本数据集;所述预设面试样本数据集中包含至少一个不具有面试标注标签的第一面试样本数据;obtaining a preset interview sample data set; the preset interview sample data set includes at least one first interview sample data without an interview label;
将所述第一面试样本数据输入至包含第一初始参数的预设识别模型中,通过所述预设 识别模型中的直接预测模块对所述第一面试样本数据进行标准标签预测,得到标准标签分布以及与所述第一面试样本数据对应的面试编码向量;Inputting the first interview sample data into a preset recognition model including the first initial parameter, and performing standard label prediction on the first interview sample data through the direct prediction module in the preset recognition model to obtain a standard label distribution and an interview coding vector corresponding to the first interview sample data;
通过所述预设识别模型中的各辅助预测模块,根据所述面试编码向量对所述第一面试样本数据进行辅助标签预测,得到与各所述辅助预测模块输出的辅助标签分布;Through each auxiliary prediction module in the preset recognition model, carry out auxiliary label prediction on the first interview sample data according to the interview coding vector, and obtain the auxiliary label distribution output by each of the auxiliary prediction modules;
根据各所述辅助标签分布与所述标准标签分布确定所述预设识别模型的总损失值;Determine the total loss value of the preset recognition model according to each of the auxiliary label distribution and the standard label distribution;
在所述总损失值未达到预设的收敛条件时,更新迭代所述预设识别模型的第一初始参数,直至所述总损失值达到所述预设的收敛条件时,将收敛之后的所述预设识别模型记录为面试实体识别模型。When the total loss value does not reach the preset convergence condition, update and iterate the first initial parameter of the preset recognition model, until the total loss value reaches the preset convergence condition, all the The aforementioned preset recognition model is recorded as the interview entity recognition model.
一个或多个存储有计算机可读指令的可读存储介质,所述计算机可读指令被一个或多个处理器执行时,使得所述一个或多个处理器执行如下步骤:One or more readable storage media storing computer-readable instructions that, when executed by one or more processors, cause the one or more processors to perform the following steps:
获取目标面试者的面试信息;所述面试信息中包含至少一个面试句子;一个所述面试句子中包含多个面试信息字词;Obtain the interview information of the target interviewee; the interview information includes at least one interview sentence; one of the interview sentences includes a plurality of interview information words;
将所述面试句子输入至面试实体识别模型中,对所述面试句子中的面试信息字词进行提取识别,得到与各所述面试信息字词对应的实体识别结果;所述面试实体识别模型是根据上述面试实体识别模型训练方法得到的;The interview sentence is input into the interview entity recognition model, and the interview information words in the interview sentence are extracted and recognized to obtain entity recognition results corresponding to the interview information words; the interview entity recognition model is Obtained according to the above interview entity recognition model training method;
将所述实体识别结果按照预设匹配规则插入预设面试信息存储模板。Insert the entity recognition result into a preset interview information storage template according to preset matching rules.
上述面试实体识别模型训练、面试信息实体提取方法、装置、设备及介质,面试实体识别模型训练方法通过获取预设面试样本数据集;所述预设面试样本数据集中包含至少一个不具有面试标注标签的第一面试样本数据;将所述第一面试样本数据输入至包含第一初始参数的预设识别模型中,通过所述预设识别模型中的直接预测模块对所述第一面试样本数据进行标准标签预测,得到标准标签分布以及与所述第一面试样本数据对应的面试编码向量;通过所述预设识别模型中的各辅助预测模块,根据所述面试编码向量对所述第一面试样本数据进行辅助标签预测,得到与各所述辅助预测模块输出的辅助标签分布;根据各所述辅助标签分布与所述标准标签分布确定所述预设识别模型的总损失值;在所述总损失值未达到预设的收敛条件时,更新迭代所述预设识别模型的第一初始参数,直至所述总损失值达到所述预设的收敛条件时,将收敛之后的所述预设识别模型记录为面试实体识别模型。The above interview entity recognition model training, interview information entity extraction method, device, equipment and medium, the interview entity recognition model training method obtains a preset interview sample data set; the preset interview sample data set includes at least one that does not have an interview label. The first interview sample data; the first interview sample data is input into a preset recognition model containing the first initial parameter, and the first interview sample data is analyzed by the direct prediction module in the preset recognition model. Standard label prediction, obtain standard label distribution and interview coding vector corresponding to the first interview sample data; through each auxiliary prediction module in the preset recognition model, according to the interview coding vector Perform auxiliary label prediction on the data to obtain auxiliary label distributions output by each of the auxiliary prediction modules; determine the total loss value of the preset recognition model according to each of the auxiliary label distributions and the standard label distribution; When the value does not reach the preset convergence condition, update and iterate the first initial parameter of the preset recognition model, until the total loss value reaches the preset convergence condition, the preset recognition model after convergence will be updated. Recorded as Interview Entity Recognition Model.
本申请中,通过将有监督学习的直接预测模块,以及无监督学习的辅助预测模块结合,以在直接预测模块的基础上,通过辅助预测模块给予模型更多不同的数据特征(如各辅助预测模块输出的辅助标签分布),提高了面试实体识别模型训练效率,并提高了训练完成的面试实体识别模型的准确率。In this application, by combining the direct prediction module of supervised learning and the auxiliary prediction module of unsupervised learning, on the basis of the direct prediction module, the auxiliary prediction module can give the model more different data features (such as each auxiliary prediction module). The auxiliary label distribution output by the module) improves the training efficiency of the interview entity recognition model and improves the accuracy of the trained interview entity recognition model.
本申请的一个或多个实施例的细节在下面的附图和描述中提出,本申请的其他特征和优点将从说明书、附图以及权利要求变得明显。The details of one or more embodiments of the application are set forth in the accompanying drawings and the description below, and other features and advantages of the application will become apparent from the description, drawings, and claims.
附图说明Description of drawings
为了更清楚地说明本申请实施例的技术方案,下面将对本申请实施例的描述中所需要使用的附图作简单地介绍,显而易见地,下面描述中的附图仅仅是本申请的一些实施例,对于本领域普通技术人员来讲,在不付出创造性劳动性的前提下,还可以根据这些附图获得其他的附图。In order to illustrate the technical solutions of the embodiments of the present application more clearly, the following briefly introduces the drawings that are used in the description of the embodiments of the present application. Obviously, the drawings in the following description are only some embodiments of the present application. , for those of ordinary skill in the art, other drawings can also be obtained from these drawings without creative labor.
图1是本申请一实施例中面试实体识别模型训练方法、面试信息实体提取方法的一应用环境示意图;1 is a schematic diagram of an application environment of an interview entity recognition model training method and an interview information entity extraction method in an embodiment of the present application;
图2是本申请一实施例中面试实体识别模型训练方法的一流程图;Fig. 2 is a flowchart of an interview entity recognition model training method in an embodiment of the present application;
图3是本申请一实施例中面试实体识别模型训练方法中步骤S30的一流程图;Fig. 3 is a flowchart of step S30 in the training method of interview entity recognition model in an embodiment of the present application;
图4是本申请一实施例中面试实体识别模型训练方法的另一流程图;Fig. 4 is another flowchart of the training method of the interview entity recognition model in an embodiment of the present application;
图5是本申请一实施例中面试信息实体提取方法的一流程图;5 is a flowchart of a method for extracting interview information entities in an embodiment of the present application;
图6是本申请一实施例中面试实体识别模型训练装置的一原理框图;6 is a schematic block diagram of an interview entity recognition model training device in an embodiment of the present application;
图7是本申请一实施例中面试实体识别模型训练装置中辅助标签预测模块的一原理框图;7 is a schematic block diagram of an auxiliary label prediction module in an interview entity recognition model training device in an embodiment of the present application;
图8是本申请一实施例中面试信息实体提取装置的一原理框图;8 is a schematic block diagram of an apparatus for extracting interview information entities in an embodiment of the present application;
图9是本申请一实施例中计算机设备的一示意图。FIG. 9 is a schematic diagram of a computer device in an embodiment of the present application.
具体实施方式Detailed ways
下面将结合本申请实施例中的附图,对本申请实施例中的技术方案进行清楚、完整地描述,显然,所描述的实施例是本申请一部分实施例,而不是全部的实施例。基于本申请中的实施例,本领域普通技术人员在没有作出创造性劳动前提下所获得的所有其他实施例,都属于本申请保护的范围。The technical solutions in the embodiments of the present application will be clearly and completely described below with reference to the accompanying drawings in the embodiments of the present application. Obviously, the described embodiments are part of the embodiments of the present application, not all of the embodiments. Based on the embodiments in the present application, all other embodiments obtained by those of ordinary skill in the art without creative work fall within the protection scope of the present application.
本申请实施例提供的面试实体识别模型训练方法,该面试实体识别模型训练方法可应用如图1所示的应用环境中。具体地,该面试实体识别模型训练方法应用在面试实体识别模型训练系统中,该面试实体识别模型训练系统包括如图1所示的客户端和服务器,客户端与服务器通过网络进行通信,用于解决监督学习对数据量的需求非常大,以及无监督学习限制模型准确率的问题。其中,客户端又称为用户端,是指与服务器相对应,为客户提供本地服务的程序。客户端可安装在但不限于各种个人计算机、笔记本电脑、智能手机、平板电脑和便携式可穿戴设备上。服务器可以用独立的服务器或者是多个服务器组成的服务器集群来实现。The interview entity recognition model training method provided by the embodiment of the present application can be applied in the application environment shown in FIG. 1 . Specifically, the interview entity recognition model training method is applied in an interview entity recognition model training system. The interview entity recognition model training system includes a client and a server as shown in FIG. 1 , and the client and the server communicate through a network for Solve the problem that supervised learning requires a large amount of data and unsupervised learning limits the accuracy of the model. Among them, the client, also known as the client, refers to the program corresponding to the server and providing local services for the client. Clients can be installed on, but not limited to, various personal computers, laptops, smartphones, tablets, and portable wearable devices. The server can be implemented as an independent server or a server cluster composed of multiple servers.
在一实施例中,如图2所示,提供一种面试实体识别模型训练方法,以该方法应用在图1中的服务器为例进行说明,包括如下步骤:In one embodiment, as shown in FIG. 2 , a method for training an interview entity recognition model is provided, and the method is applied to the server in FIG. 1 as an example for description, including the following steps:
S10:获取预设面试样本数据集;所述预设面试样本数据集中包含至少一个不具有面试标注标签的第一面试样本数据;S10: Obtain a preset interview sample data set; the preset interview sample data set includes at least one first interview sample data without an interview label;
可以理解地,第一面试样本数据为不具有预先通过人工标注的面试标注标签的数据;一般地,在有监督学习中需要大量的人工标注数据进行模型训练学习,但是人工标注数据需求量很大,通过人工进行标注的方法浪费时间,且无法输出庞大的标注数据,因此本申请需要解决的其中一个问题就是缺乏有标注数据的情况下,如何对模型进行更加精确,快速的训练学习。进一步地,该第一面试样本数据可以根据不同场景进行选取,示例性地,第一面试样本数据可以为面试者的自我介绍,亦或者面试者的简历;此外,本申请还可以应用在如电影编辑场景下,该第一面试样本数据可以替换为为电影剧本中的句子。Understandably, the first interview sample data is data that does not have pre-marked interview labels; generally, a large amount of manually labeled data is required for model training and learning in supervised learning, but the demand for manually labeled data is large. , the manual labeling method wastes time and cannot output huge labeled data. Therefore, one of the problems to be solved in this application is how to train and learn the model more accurately and quickly in the absence of labeled data. Further, the first interview sample data can be selected according to different scenarios. Exemplarily, the first interview sample data can be the interviewee's self-introduction, or the interviewer's resume; In the editing scenario, the first interview sample data can be replaced with sentences in the movie script.
S20:将所述第一面试样本数据输入至包含第一初始参数的预设识别模型中,通过所述预设识别模型中的直接预测模块对所述第一面试样本数据进行标准标签预测,得到标准标签分布以及与所述第一面试样本数据对应的面试编码向量;S20: Input the first interview sample data into a preset recognition model including the first initial parameter, and perform standard label prediction on the first interview sample data through the direct prediction module in the preset recognition model, to obtain Standard label distribution and interview coding vector corresponding to the first interview sample data;
可以理解地,在本申请中,预设识别模型是结合了有监督学习以及无监督学习形成的半监督学习模型;直接预测模块是通过少量的具有面试标注标签的数据进行训练得到的,也即直接预测模块是训练完成的模块,进而通过直接预测模块对不具有面试标注标签的第一面试样本数据进行标准标签预测时,可以不用额外训练一个预测模块,提高了模型训练的效率。Understandably, in this application, the preset recognition model is a semi-supervised learning model formed by combining supervised learning and unsupervised learning; the direct prediction module is obtained by training a small amount of data with interview labels, that is, The direct prediction module is a module that has been trained. When the standard label prediction is performed on the first interview sample data that does not have an interview label through the direct prediction module, it is not necessary to train an additional prediction module, which improves the efficiency of model training.
进一步地,在将第一面试样本数据输入至包含第一初始参数的预设识别模型之后,该第一面试样本数据作为直接预测模块的输入,该直接预测模块中包含一个双向循环神经网络编码器,该双向循环神经网络编码器用于对第一面试样本数据进行向量编码,进而得到与第一面试样本数据对应的面试编码向量;该直接预测模块中还包含一个通过具有面试标注标签训练的标注分类器,进而在通过双向循环神经网络编码器对第一面试样本数据进行向量编码得到面试编码向量后,通过标注分类器对面试编码向量进行直接标签预测,得到与所述第一面试样本数据对应的标准标签分布。Further, after inputting the first interview sample data into the preset recognition model including the first initial parameter, the first interview sample data is used as the input of the direct prediction module, which includes a bidirectional recurrent neural network encoder. , the bidirectional cyclic neural network encoder is used for vector encoding the first interview sample data, and then obtains the interview encoding vector corresponding to the first interview sample data; the direct prediction module also includes an annotation classification trained with interview annotation labels. Then, after vector encoding the first interview sample data through the bidirectional cyclic neural network encoder to obtain the interview encoding vector, the labeling classifier is used to perform direct label prediction on the interview encoding vector to obtain the corresponding first interview sample data. Standard label distribution.
S30:通过所述预设识别模型中的各辅助预测模块,根据所述面试编码向量对所述第一面试样本数据进行辅助标签预测,得到与各所述辅助预测模块输出的辅助标签分布;S30: Through each auxiliary prediction module in the preset recognition model, perform auxiliary label prediction on the first interview sample data according to the interview coding vector, and obtain the auxiliary label distribution output by each of the auxiliary prediction modules;
可以理解地,辅助预测模块指的是根据不同的字词组合对某个字词进行实体预测的模块,该辅助预测模块用于与直接预测模块进行结合形成半监督模式,对如第一标注数据等不具有面试标注标签的数据进行实体预测;需要说明的是,为了尽可能多的提取第一面试样本数据中各个字词的表征数据,每一个辅助预测模块提取第一面试样本数据的特征均是不同的,也即每一个辅助预测模块对第一面试样本数据中字词的实体判别的依据是不一样的,进而提高模型对实体识别的准确率;示例性地,如通过目标预测字词前三个字词对其进行实体预测,亦或者通过目标预测字词后三个字词对其进行实体预测等。Understandably, the auxiliary prediction module refers to a module that performs entity prediction on a certain word according to different word combinations. The auxiliary prediction module is used to combine with the direct prediction module to form a semi-supervised mode. It should be noted that, in order to extract as much representation data of each word in the first interview sample data as possible, the features of the first interview sample data extracted by each auxiliary prediction module are all are different, that is, the basis for each auxiliary prediction module to discriminate the entities of the words in the first interview sample data is different, thereby improving the accuracy of the model's entity recognition; exemplarily, such as predicting words through the target Entity prediction is performed on the first three words, or entity prediction is performed on the last three words of the target prediction word, etc.
具体地,在将所述第一面试样本数据输入至包含第一初始参数的预设识别模型中,通过所述预设识别模型中的直接预测模块对所述第一面试样本数据进行标准标签预测,得到标准标签分布以及与所述第一面试样本数据对应的面试编码向量之后,通过预设识别模型中的各辅助预测模块,根据面试编码向量对第一面试样本数据进行不同视图的辅助标签预测,可以理解地,上述说明中已经指出各个辅助预测模块对第一面试样本数据中字词的实体判别的依据是不一样的,也即各个辅助预测模块是以不同的字词视图进行辅助标签预测,进而输出对第一面试样本数据中各字词的实体预测结果,也即辅助标签分布。Specifically, in inputting the first interview sample data into a preset recognition model including first initial parameters, standard label prediction is performed on the first interview sample data through the direct prediction module in the preset recognition model , after obtaining the standard label distribution and the interview coding vector corresponding to the first interview sample data, through each auxiliary prediction module in the preset recognition model, the first interview sample data is subjected to different views of auxiliary label prediction according to the interview coding vector It can be understood that the above description has pointed out that the basis of each auxiliary prediction module for the entity discrimination of words in the first interview sample data is different, that is, each auxiliary prediction module uses different word views to predict auxiliary labels. , and then output the entity prediction result for each word in the first interview sample data, that is, the auxiliary label distribution.
在一具体实施例中,如图3所示,步骤S30中,包括:In a specific embodiment, as shown in FIG. 3 , step S30 includes:
S301:获取所述面试编码向量中的至少两个第二正向编码向量以及至少两个第二反向编码向量;所述第二正向编码是指对各所述未标注样本字词按照正向顺序进行编码得到;所述第二反向编码是指对各所述未标注样本字词按照反向顺序进行编码得到;S301: Obtain at least two second forward coding vectors and at least two second reverse coding vectors in the interview coding vectors; the second forward coding refers to performing the forward coding on each of the unlabeled sample words according to the normal The second reverse encoding refers to encoding each of the unlabeled sample words in reverse order to obtain;
可以理解地,未标注样本字词指的是对第一面试样本数据进行分词后得到的各个分词。一个辅助预测模块关联一个第二正向编码向量。Understandably, the unlabeled sample words refer to each segmented word obtained after segmenting the first interview sample data. An auxiliary prediction module is associated with a second forward coding vector.
进一步地,通过预设识别模型中的直接预测模块中的编码器对第一面试样本数据进行编码处理后,可以理解地,该编码器为双向循环神经网络编码器,因此输出的编码向量中包含将第一面试样本数据中各未标注样本字词按照正向顺序进行编码得到的第二正向编码向量(也即从第一面试样本数据中第一个未标注样本字词起,从前往后编码),以及将第一面试样本数据中各未标注样板字词按照反向顺序进行编码得到的第二反向编码向量(也即从第一面试样本数据中最后一个未标注样本字词起,从后往前编码)。Further, after encoding the first interview sample data by the encoder in the direct prediction module in the preset recognition model, it can be understood that the encoder is a bidirectional cyclic neural network encoder, so the output encoding vector contains The second forward coding vector obtained by encoding the unlabeled sample words in the first interview sample data in the forward order (that is, starting from the first unlabeled sample word in the first interview sample data, from front to back coding), and the second reverse coding vector obtained by encoding each unlabeled sample word in the first interview sample data in reverse order (that is, starting from the last unlabeled sample word in the first interview sample data, Encoding from back to front).
示例性地,在本实施例中获取样本向量编码中至少两个第二正向编码向量指的是,第二正向编码向量可以为由在该未标注样本字词前t个未标注样本字词组成的编码向量序列;第二正向编码向量还可以为由在该未标注样本字词前t‐1个未标注样本字词组成的编码向量序列;同理,至少两个第二反向编码向量指的是,第二反向编码向量可以为由未标注样本字词后t个未标注样本字词组成的编码向量序列;第二反向编码向量还可以为由未标注样本字词后t+1个未标注样本字词组成的编码向量序列。Exemplarily, in this embodiment, obtaining at least two second forward coding vectors in the coding of the sample vector means that the second forward coding vector may be the first t unlabeled sample words in the unlabeled sample word. A coding vector sequence composed of words; the second forward coding vector can also be a coding vector sequence composed of t-1 unlabeled sample words before the unlabeled sample word; similarly, at least two second reverse The coding vector refers to that the second reverse coding vector may be a coding vector sequence consisting of t unlabeled sample words after the unlabeled sample word; A sequence of encoded vectors consisting of t+1 unlabeled sample words.
S302:根据各所述第二正向编码向量确定与各所述未标注样本字词对应的各正向辅助标签分布;同时,根据所述第二反向编码向量确定与各所述未标注样本字词对应的反向辅助标签分布。S302: Determine the distribution of forward auxiliary labels corresponding to each of the unlabeled sample words according to each of the second forward coding vectors; Reverse auxiliary label distribution for words.
可以理解地,在获取所述面试编码向量中的至少两个第二正向编码向量以及至少两个第二反向编码向量之后,通过对应的第二正向量编码中的各未标注样本字词进行实体预测,得到与各第二正向编码向量对应的正向辅助标签分布,也即一个第二正向编码向量输入至辅助预测模块之后,会输出一个与其相对应的正向辅助标签分布。同理,通过对应的第二反向向量编码中的各未标注样本字词进行实体预测,得到与各第二反向编码向量对应的反向辅助标签分布。Understandably, after obtaining at least two second forward coding vectors and at least two second reverse coding vectors in the interview coding vector, each unlabeled sample word in the corresponding second forward vector coding is Perform entity prediction to obtain the forward auxiliary label distribution corresponding to each second forward coding vector, that is, after a second forward coding vector is input to the auxiliary prediction module, a corresponding forward auxiliary label distribution will be output. Similarly, entity prediction is performed by each unlabeled sample word in the corresponding second reverse vector encoding, and the reverse auxiliary label distribution corresponding to each second reverse encoding vector is obtained.
示例性地,假设需要对第一面试样本数据中未标注样本字词A进行实体预测,且第二正向编码向量可以为由在该未标注样本字词A前t个未标注样本字词组成的编码向量序列, 因此在该辅助预测模块中通过前t个未标注样本字词对未标注样本字词A进行实体预测;若第二正向编码向量为由在该未标注样本字词A前t‐1个未标注样本字词组成的编码向量序列,则在另一个辅助预测模块中通过前t‐1个未标注样本字词对未标注样本字词A进行实体预测。可以理解地,上述两个第二正向编码向量对未标注样本字词A进行实体预测后输出的正向辅助标签分布是不同的,由于通过前t个未标注样本字词对未标注样本字词A进行实体预测时,可以触及该未标注样本字词A;而通过前t‐1个未标注样本字词对未标注样本字词A进行实体预测时,在未标注样本字词A之前还存在一个其它未标注样本字词,此时无法触及该未标注样本字词A,进而形成辅助预测模块在对未标注样本字词A进行实体预测时的视图差异,第二反向编码向量同理,在此不再赘述。Exemplarily, it is assumed that entity prediction needs to be performed on the unlabeled sample word A in the first interview sample data, and the second forward encoding vector may be composed of t unlabeled sample words before the unlabeled sample word A. Therefore, in the auxiliary prediction module, entity prediction is performed on the unlabeled sample word A through the first t unlabeled sample words; The encoding vector sequence composed of t-1 unlabeled sample words, in another auxiliary prediction module, entity prediction is performed on the unlabeled sample word A through the first t-1 unlabeled sample words. It can be understood that the distribution of the forward auxiliary labels output after the entity prediction of the unlabeled sample word A by the above-mentioned two second forward coding vectors is different, because the unlabeled sample words are compared to the unlabeled sample words through the first t unlabeled sample words. When the word A is used for entity prediction, the unlabeled sample word A can be touched; and when the unlabeled sample word A is predicted by the first t-1 unlabeled sample words, the unlabeled sample word A is also before the unlabeled sample word A. There is another unlabeled sample word, and the unlabeled sample word A cannot be touched at this time, thereby forming the view difference when the auxiliary prediction module performs entity prediction on the unlabeled sample word A. The same is true for the second reverse encoding vector , and will not be repeated here.
S40:根据各所述辅助标签分布与所述标准标签分布确定所述预设识别模型的总损失值;S40: Determine the total loss value of the preset recognition model according to the distribution of each auxiliary label and the distribution of the standard label;
可以理解地,在通过所述预设识别模型中的各辅助预测模块,根据所述编码向量对所述第一面试样本数据进行辅助标签预测,得到与各所述辅助预测模块输出的辅助标签分布之后,确定各辅助标签分布与标准标签分布之间的KL(Kullback–Leibler divergence,相对熵)散度,具体地可以根据如下表达式确定:It can be understood that, through each auxiliary prediction module in the preset recognition model, auxiliary label prediction is performed on the first interview sample data according to the coding vector, and the auxiliary label distribution output by each auxiliary prediction module is obtained. After that, determine the KL (Kullback–Leibler divergence, relative entropy) divergence between each auxiliary label distribution and the standard label distribution, which can be specifically determined according to the following expression:
Figure PCTCN2021096583-appb-000001
Figure PCTCN2021096583-appb-000001
其中,D KL(p||q)指的是辅助标签分布与面试标注标签分布之间的KL散度;p(x i)表征的是第一面试样本数据中第i个未标注样本字词对应的辅助预测模块输出的辅助标签分布;q(x i)表征的是与p(x i)的未标注样本字词对应的标准标签分布。 Among them, D KL (p||q) refers to the KL divergence between the auxiliary label distribution and the interview label distribution; p(x i ) represents the ith unlabeled sample word in the first interview sample data The auxiliary label distribution output by the corresponding auxiliary prediction module; q(x i ) represents the standard label distribution corresponding to the unlabeled sample words of p(x i ).
进一步地,通过下述表达式确定预设识别模型的总损失值:Further, the total loss value of the preset recognition model is determined by the following expression:
Figure PCTCN2021096583-appb-000002
Figure PCTCN2021096583-appb-000002
其中,L VCT(θ)为预设识别模型的总损失值;|D ul|为预设面试样本数据集中第一面试样本数据的个数;k为预设识别模型中辅助预测模块的个数;q θ(y|x i)为第θ个第一面试样本数据中第i个未标注样本字词对应的标准标签分布;
Figure PCTCN2021096583-appb-000003
为第θ个第一面试样本数据中第i个未标注样本字词的第j个辅助预测模块输出的辅助标签分布;
Figure PCTCN2021096583-appb-000004
为第θ个第一面试样本数据中第i个未标注样本字词的各辅助标签分布与标准标签分布之间的KL散度。
Among them, L VCT (θ) is the total loss value of the preset recognition model; |D ul | is the number of the first interview sample data in the preset interview sample data set; k is the number of auxiliary prediction modules in the preset recognition model ; q θ (y|x i ) is the standard label distribution corresponding to the i-th unlabeled sample word in the θ-th first interview sample data;
Figure PCTCN2021096583-appb-000003
is the auxiliary label distribution output by the jth auxiliary prediction module of the ith unlabeled sample word in the θth first interview sample data;
Figure PCTCN2021096583-appb-000004
is the KL divergence between each auxiliary label distribution and the standard label distribution of the i-th unlabeled sample word in the θ-th first interview sample data.
S50:在所述总损失值未达到预设的收敛条件时,更新迭代所述预设识别模型的第一初始参数,直至所述总损失值达到所述预设的收敛条件时,将收敛之后的所述预设识别模型记录为面试实体识别模型。S50: When the total loss value does not reach the preset convergence condition, update and iterate the first initial parameter of the preset recognition model, until the total loss value reaches the preset convergence condition, after the convergence The preset recognition model is recorded as the interview entity recognition model.
可以理解地,该收敛条件可以为总损失值小于设定阈值的条件,也即在总损失值小于设定阈值时,停止训练;收敛条件还可以为总损失值经过了10000次计算后值为很小且不会再下降的条件,也即总损失值经过10000次计算后值很小且不会下降时,停止训练,将收敛之后的所述预设识别模型记录为面试实体识别模型。Understandably, the convergence condition can be the condition that the total loss value is less than the set threshold, that is, when the total loss value is less than the set threshold, the training is stopped; the convergence condition can also be that the total loss value after 10,000 calculations is If the condition is very small and will not decrease again, that is, when the total loss value is small and will not decrease after 10,000 calculations, the training is stopped, and the preset recognition model after convergence is recorded as the interview entity recognition model.
进一步地,根据所述第一面试样本数据对应的根据各所述辅助标签分布与所述标准标 签分布确定所述预设识别模型的总损失值之后,在总损失值未达到预设的收敛条件时,根据该总损失值调整预设识别模型的第一初始参数,并将该第一面试样本数据重新输入至调整初始参数后的预设识别模型中,以在该第一面试样本数据对应的总损失值达到预设的收敛条件时,选取预设面试样本数据集中另一仅第一面试样本数据,并执行上述步骤S10至S50,并得到与该第一面试样本数据对应的预测损失值,并在该总损失值未达到预设的收敛条件时,根据该总损失值再次调整预设识别模型的第一初始参数,使得该第一面试样本数据对应的总损失值达到预设的收敛条件。Further, after determining the total loss value of the preset recognition model according to the distribution of each auxiliary label and the standard label distribution corresponding to the first interview sample data, after the total loss value does not reach the preset convergence condition , adjust the first initial parameters of the preset recognition model according to the total loss value, and re-input the first interview sample data into the preset recognition model after adjusting the initial parameters, so that the first interview sample data corresponds to the When the total loss value reaches the preset convergence condition, select another first interview sample data in the preset interview sample data set, and execute the above steps S10 to S50, and obtain the predicted loss value corresponding to the first interview sample data, And when the total loss value does not reach the preset convergence condition, adjust the first initial parameter of the preset recognition model again according to the total loss value, so that the total loss value corresponding to the first interview sample data reaches the preset convergence condition .
如此,在通过预设面试样本数据集中所有第一面试样本数据对预设识别模型进行训练之后,使得预设识别模型输出的结果可以不断向准确地结果靠拢,让识别准确率越来越高,直至所有第一面试样本数据对应的总损失值均达到预设的收敛条件时,将收敛之后的所述预设识别模型记录为面试实体识别模型。In this way, after training the preset recognition model with all the first interview sample data in the preset interview sample data set, the output results of the preset recognition model can be continuously approached to the accurate results, so that the recognition accuracy rate is getting higher and higher. Until the total loss value corresponding to all the first interview sample data reaches the preset convergence condition, the preset recognition model after convergence is recorded as the interview entity recognition model.
在本实施例中,通过将有监督学习的直接预测模块,以及无监督学习的辅助预测模块结合,以在直接预测模块的基础上,通过辅助预测模块给予模型更多不同的数据特征(如各辅助预测模块输出的辅助标签分布),提高了面试实体识别模型训练效率,并提高了训练完成的面试实体识别模型的准确率。In this embodiment, by combining the direct prediction module of supervised learning and the auxiliary prediction module of unsupervised learning, on the basis of the direct prediction module, the auxiliary prediction module can give the model more different data features (such as each The auxiliary label distribution output by the auxiliary prediction module) improves the training efficiency of the interview entity recognition model and improves the accuracy of the trained interview entity recognition model.
在另一具体实施例中,为了保证上述实施例中的面试实体识别模型的私密以及安全性,可以将面试实体识别模型存储在区块链中。其中,区块链(Blockchain),是由区块(Block)形成的加密的、链式的交易的存储结构。In another specific embodiment, in order to ensure the privacy and security of the interview entity recognition model in the above embodiment, the interview entity recognition model may be stored in the blockchain. Among them, Blockchain is a storage structure of encrypted and chained transactions formed by blocks.
例如,每个区块的头部既可以包括区块中所有交易的哈希值,同时也包含前一个区块中所有交易的哈希值,从而基于哈希值实现区块中交易的防篡改和防伪造;新产生的交易被填充到区块并经过区块链网络中节点的共识后,会被追加到区块链的尾部从而形成链式的增长。For example, the header of each block can include not only the hash values of all transactions in the block, but also the hash values of all transactions in the previous block, so that the transactions in the block can be tamper-proof based on the hash value. And anti-counterfeiting; the newly generated transaction is filled into the block and after the consensus of the nodes in the blockchain network, it will be appended to the end of the blockchain to form a chain growth.
在一实施例中,如图4所示,所述预设面试样本数据集中还包含至少一个具有所述面试标注标签的第二面试样本数据;步骤S20之前,也即所述通过所述预设识别模型中的预设编码器对所述第一面试样本数据进行标准标签预测之前,还包括:In one embodiment, as shown in FIG. 4 , the preset interview sample data set also includes at least one second interview sample data with the interview label; Before the preset encoder in the recognition model performs standard label prediction on the first interview sample data, the method further includes:
S01:将所述第二面试样本数据输入至所述预设识别模型中,通过所述预设识别模型中包含第二初始参数的预设预测模块对所述第二面试样本数据进行直接标签预测,得到与所述第二面试样本数据对应的直接预测标签;S01: Input the second interview sample data into the preset recognition model, and perform direct label prediction on the second interview sample data through a preset prediction module that includes a second initial parameter in the preset recognition model , obtain the direct prediction label corresponding to the second interview sample data;
可以理解地,面试标注标签指的是预先对第二面试样本数据中各字词进行人工标注得到的,示例性地,“我在北京大学上学”,则会预先将“北京大学”标注为学校名称实体。It is understandable that the interview label refers to the manual labeling of each word in the second interview sample data in advance. For example, "I go to school in Peking University", "Peking University" will be marked as a school in advance. Name entity.
进一步地,在获取预设面试样本数据集之后,将包含面试标注标签的第二面试样本数据输入至预设识别模型中,通过预设识别模型中包含第二参数的预设预测模块对第二面试样本数据进行直接标签预测,得到与所述第二面试样本数据中各字词对应的直接预测标签。Further, after obtaining the preset interview sample data set, input the second interview sample data including the interview label into the preset recognition model, and use the preset prediction module including the second parameter in the preset recognition model Perform direct label prediction on the interview sample data to obtain direct predicted labels corresponding to each word in the second interview sample data.
在一具体实施例中,步骤S01中包括:In a specific embodiment, step S01 includes:
对所述第二面试样本数据进行分词处理,得到与所述第二面试样本数据对应各标注样本字词;Perform word segmentation processing on the second interview sample data to obtain each marked sample word corresponding to the second interview sample data;
示例性地,可以通过结巴分词方法对第二面试样本数据进行分词处理,得到与第二面试样本数据对应的各标注样本字词。Exemplarily, word segmentation processing may be performed on the second interview sample data through a stammering word segmentation method to obtain each marked sample word corresponding to the second interview sample data.
通过所述预设预测模块中的编码器对各所述标注样本字词进行编码处理,得到第一正向编码向量以及第一反向编码向量;所述第一正向编码是指对各所述标注样本字词按照正向顺序进行编码得到;所述第一反向编码是指对各所述标注样本字词按照反向顺序进行编码得到;The encoder in the preset prediction module performs encoding processing on each of the labeled sample words to obtain a first forward encoding vector and a first reverse encoding vector; the first forward encoding refers to The labeled sample words are obtained by encoding in a forward order; the first reverse encoding refers to encoding each of the labeled sample words in a reverse order;
其中,编码器是双向循环神经网络编码器,因此输出的编码向量中包含将第二面试样本数据中各标注样本字词按照正向顺序进行编码得到的第一正向编码向量,以及将第二面试样本数据中各标注样本字词按照反向顺序进行编码得到的第一反向编码向量。The encoder is a bidirectional cyclic neural network encoder, so the output encoding vector includes the first forward encoding vector obtained by encoding the labeled sample words in the second interview sample data in the forward order, and the second The first reverse encoding vector obtained by encoding each labeled sample word in the interview sample data in reverse order.
根据所述第一正向编码向量以及所述第一反向编码向量,通过所述预设预测模块中的标注分类器对各所述标注样本字词进行标签分类,得到与各所述标注样本字词对应的直接预测标签。According to the first forward coding vector and the first reverse coding vector, the labeling classifier in the preset prediction module is used to label each of the labelled sample words, and obtain the same label as each labelled sample. The direct prediction label corresponding to the term.
可以理解地,在通过所述预设预测模块中的编码器对各所述标注样本字词进行编码处理,得到第一正向编码向量以及第一反向编码向量之后,预设预测模块中的标注分类器则会根据第一正向编码向量以及第一反向编码向量,对各标注样本字词进行标签分类,得到与各标注样本字词对应的直接预测标签。示例性地,假设第二面试样本数据为“我在北京大学学习”,其中“北京大学”被人工标注为学校名称实体,进而在将该第二面试样本数据输入至预设预测模块之后,对该第二面试样本数据中各标注字词(“我”、“在”、“北京大学”、“学习”、)进行编码处理,得到从“我”开始,从前往后编码的第一正向编码向量,以及从“北京大学”开始,从后往前编码的第二反向编码向量,以根据第二正向编码向量以及第二反向编码向量对“北京大学”进行标签分类,得到与“北京大学”对应的直接预测标签。It can be understood that after the encoder in the preset prediction module performs encoding processing on each of the labeled sample words to obtain the first forward encoding vector and the first reverse encoding vector, the preset prediction module The labeling classifier performs label classification for each labeling sample word according to the first forward coding vector and the first reverse coding vector, and obtains a direct prediction label corresponding to each labeling sample word. Exemplarily, assume that the second interview sample data is "I study at Peking University", where "Peking University" is manually marked as a school name entity, and then after the second interview sample data is input into the preset prediction module, the Each marked word (“I”, “Zai”, “Peking University”, “Learning”, etc.) in the second interview sample data is coded to obtain the first forward coding starting from “I” and coded from front to back The encoding vector, and the second reverse encoding vector encoded from the back to the front starting from "Peking University", to classify "Peking University" according to the second forward encoding vector and the second reverse encoding vector. The direct prediction label corresponding to "Peking University".
S02:根据所述直接预测标签与所述面试标注标签确定所述预设编码器的预测损失值;S02: Determine the prediction loss value of the preset encoder according to the direct prediction label and the interview label;
可以理解地,在根据所述第一正向编码向量以及所述第一反向编码向量,通过所述预设预测模块中的标注分类器对各所述标注样本字词进行标签分类,得到与各所述标注样本字词对应的直接预测标签之后,将各标注样本字词对应的直接预测标签,与各标注样本字词对应的面试标注标签进行匹配,进而确定预设编码器的预测损失值。It can be understood that, according to the first forward coding vector and the first reverse coding vector, the labeling classifier in the preset prediction module performs label classification on each of the labeled sample words, and obtains the same After the direct prediction label corresponding to each labeled sample word, the direct prediction label corresponding to each labeled sample word is matched with the interview labeled label corresponding to each labeled sample word, and then the prediction loss value of the preset encoder is determined. .
具体地,所述面试标注标签中包含多个样本实体标签;步骤S02中,包括:Specifically, the interview label includes a plurality of sample entity labels; in step S02, including:
(1)获取与各所述标注样本字词对应的所述样本实体标签;(1) Obtain the sample entity labels corresponding to each of the labeled sample words;
(2)根据与同一个标注样本字词对应的所述样本实体标签以及所述直接预测标签,确定与该标注样本字词对应的标签损失值;(2) According to the sample entity label and the direct prediction label corresponding to the same labeled sample word, determine the label loss value corresponding to the labeled sample word;
(3)根据与各所述标注样本字词对应的标签损失值,通过交差熵损失函数确定所述预测损失值。(3) According to the label loss value corresponding to each labeled sample word, the predicted loss value is determined by the cross entropy loss function.
可以理解地,样本实体标签即为一个第二面试样本数据中存在标注的各个标注样本字词对应的标签,也即一个第二面试样本数据中可能存在两个及两个以上被标注的字词。因此,在将所述第二面试样本数据输入至所述预设识别模型中,通过所述预设识别模型中包含第二初始参数的预设预测模块对所述第二面试样本数据进行直接标签预测,得到与所述第二面试样本数据对应的直接预测标签之后,获取与各所述标注样本字词对应的所述样本实体标签,根据与同一个标注样本字词对应的所述样本实体标签以及所述直接预测标签,确定与该标注样本字词对应的标签损失值;根据与各所述标注样本字词对应的标签损失值,通过交差熵损失函数确定所述预测损失值。Understandably, the sample entity label is the label corresponding to each labeled sample word that exists in a second interview sample data, that is, there may be two or more labeled words in a second interview sample data. . Therefore, when the second interview sample data is input into the preset recognition model, the second interview sample data is directly labeled by the preset prediction module including the second initial parameter in the preset recognition model Predict, after obtaining the direct prediction label corresponding to the second interview sample data, obtain the sample entity label corresponding to each of the labeled sample words, according to the sample entity label corresponding to the same labeled sample word and the direct prediction label, to determine the label loss value corresponding to the labeled sample word; according to the label loss value corresponding to each labeled sample word, determine the predicted loss value through the cross entropy loss function.
进一步地,可以根据下述表达式确定预测损失值:Further, the predicted loss value can be determined according to the following expression:
Figure PCTCN2021096583-appb-000005
Figure PCTCN2021096583-appb-000005
其中,L sup(β)为预测损失值;|D l|为所有第二面试样本数据的个数;
Figure PCTCN2021096583-appb-000006
指的是第β个第二面试样本数据中第δ个标注样本字词对应的样本实体标签;p β(b|a δ)指的是第β个第二面试样本数据中第δ个标注样本字词对应的直接预测标签;CE()为交叉熵损失函数。
Among them, L sup (β) is the prediction loss value; |D l | is the number of all the second interview sample data;
Figure PCTCN2021096583-appb-000006
refers to the sample entity label corresponding to the δth labeled sample word in the βth second interview sample data; p β (b|a δ ) refers to the δth labeled sample in the βth second interview sample data The direct prediction label corresponding to the word; CE() is the cross entropy loss function.
可以理解地,针对于整个面试实体识别模型而言,其在训练过程中首先通过具有面试标注标签的第二面试样本数据训练直接预测模块,进而通过直接预测模块对不具有面试标注标签的第一面试样本数据进行标准标签预测,以及通过各辅助预测模块对第一面试样本数据进行辅助标签预测,进而表征本申请中的面试实体识别模型的整体损失值包含两个部分,也即步骤S02中的预测损失值以及步骤S40中的总损失值的叠加。Understandably, for the entire interview entity recognition model, in the training process, it first trains the direct prediction module through the second interview sample data with the interview label, and then uses the direct prediction module to analyze the first interview without the interview label through the direct prediction module. The standard label prediction is performed on the interview sample data, and the auxiliary label prediction is performed on the first interview sample data through each auxiliary prediction module, thereby characterizing the overall loss value of the interview entity recognition model in this application. A superposition of the predicted loss value and the total loss value in step S40.
S03:在所述预测损失值未达到预设的收敛条件时,更新迭代所述预设预设预测模块的第二初始参数,直至所述预测损失值达到所述预设的收敛条件时,将收敛之后的所述预设预测模块记录为所述直接预测模块。S03: When the predicted loss value does not reach the preset convergence condition, update and iterate the second initial parameter of the preset preset prediction module, until the predicted loss value reaches the preset convergence condition, change the The preset prediction module after convergence is recorded as the direct prediction module.
可以理解地,该收敛条件可以为预测损失值小于设定阈值的条件,也即在预测损失值小于设定阈值时,停止训练;收敛条件还可以为预测损失值经过了10000次计算后值为很小且不会再下降的条件,也即预测损失值经过10000次计算后值很小且不会下降时,停止训练,并将收敛之后的所述预设预测模块记录为所述直接预测模块。Understandably, the convergence condition can be the condition that the predicted loss value is less than the set threshold, that is, when the predicted loss value is less than the set threshold, the training is stopped; the convergence condition can also be that the predicted loss value after 10,000 calculations is The condition that is very small and will not decrease again, that is, when the predicted loss value is small and will not decrease after 10,000 calculations, stop training, and record the preset prediction module after convergence as the direct prediction module .
进一步地,根据所述第二面试样本数据的直接预测标签与所述面试标注标签确定所述预设预测模块的预测损失值之后,在预测损失值未达到预设的收敛条件时,根据该预测损失值调整预设预测模块的第二初始参数,并将该第二面试样本数据重新输入至调整初始参数后的预设预测模块中,以在该第二面试样本数据对应的预测损失值达到预设的收敛条件时,选取预设面试样本数据集中另一第二面试样本数据,并执行上述步骤S01至S03,并得到与该第二面试样本数据对应的预测损失值,并在该预测损失值未达到预设的收敛条件时,根据该预测损失值再次调整预设预测模块的第二初始参数,使得该第二面试样本数据对应的预测损失值达到预设的收敛条件。Further, after determining the predicted loss value of the preset prediction module according to the direct prediction label of the second interview sample data and the interview label, when the predicted loss value does not reach the preset convergence condition, according to the prediction The loss value adjusts the second initial parameter of the preset prediction module, and re-inputs the second interview sample data into the preset prediction module after adjusting the initial parameters, so that the predicted loss value corresponding to the second interview sample data reaches the preset value. When the convergence condition is set, select another second interview sample data in the preset interview sample data set, and perform the above steps S01 to S03, and obtain the predicted loss value corresponding to the second interview sample data, and in the predicted loss value When the preset convergence condition is not reached, the second initial parameter of the preset prediction module is adjusted again according to the predicted loss value, so that the predicted loss value corresponding to the second interview sample data reaches the preset convergence condition.
如此,在通过预设面试样本数据集中所有第二面试样本数据对预设预测模块进行训练之后,使得预设预测模块输出的结果可以不断向准确地结果靠拢,让识别准确率越来越高,直至所有第二面试样本数据对应的预测损失值均达到预设的收敛条件时,将收敛之后的所述预设预测模块记录为所述直接预测模块。In this way, after the preset prediction module is trained by using all the second interview sample data in the preset interview sample data set, the results output by the preset prediction module can be continuously approached to the accurate results, so that the recognition accuracy rate is getting higher and higher. Until all the prediction loss values corresponding to the second interview sample data reach the preset convergence condition, the preset prediction module after convergence is recorded as the direct prediction module.
在一实施例中,如图5所示,本申请还提出一种面试信息实体提取方法,以该方法应用在图1中的服务器为例进行说明,包括如下步骤:In one embodiment, as shown in FIG. 5 , the present application also proposes a method for extracting interview information entities, which is described by taking the method applied to the server in FIG. 1 as an example, including the following steps:
S60:获取目标面试者的面试信息;所述面试信息中包含至少一个面试句子;一个所述面试句子中包含多个面试信息字词;S60: Obtain interview information of the target interviewee; the interview information includes at least one interview sentence; one of the interview sentences includes a plurality of interview information words;
可以理解地,面试信息可以为目标面试者提供的纸质简历中的信息,也可以为目标面试者在自我介绍过程中,通过录取目标面试者的语音信息并转换为文字信息之后得到的。Understandably, the interview information can be the information in the paper resume provided by the target interviewer, or it can be obtained by the target interviewee by taking the voice information of the target interviewer and converting it into text information during the self-introduction process.
S70:将所述面试句子输入至面试实体识别模型中,对所述面试句子中的面试信息字词进行提取识别,得到与各所述面试信息字词对应的实体识别结果;所述面试实体识别模型是根据如上述面试实体识别模型训练方法得到的;S70: Input the interview sentence into the interview entity recognition model, extract and identify the interview information words in the interview sentence, and obtain entity recognition results corresponding to the interview information words; the interview entity recognition The model is obtained according to the above-mentioned interview entity recognition model training method;
具体地,在获取目标面试者的面试信息之后,将面试信息以句号结尾对面试信息进行拆分,以得到面试句子,并将面试句子输入至面试实体识别模型中,对面试句子中的面试信息字词进行实体识别,以确定各面试信息字词对应的实体识别结果。可以理解地,实体识别结果中包含具体实体结果,也即识别出面试信息字词具体对应的实体名称,如学校名称实体、姓名名称实体等;实体识别结果中还包括非实体识别结果,如“我”、“是”等面试信息字词为非实体。Specifically, after obtaining the interview information of the target interviewee, split the interview information with a period at the end of the interview information to obtain an interview sentence, input the interview sentence into the interview entity recognition model, and analyze the interview information in the interview sentence. Entity recognition is performed on the words to determine the entity recognition result corresponding to each interview information word. Understandably, the entity recognition results include specific entity results, that is, the entity names corresponding to the interview information words are identified, such as school name entities, name entities, etc.; entity recognition results also include non-entity recognition results, such as " Interview information words such as "I" and "Yes" are non-entity.
S80:将所述实体识别结果按照预设匹配规则插入预设面试信息存储模板。S80: Insert the entity recognition result into a preset interview information storage template according to a preset matching rule.
具体地,在将所述面试句子输入至面试实体识别模型中,对所述面试句子中的面试信息字词进行提取识别,得到与各所述面试信息字词对应的实体识别结果之后,将实体识别结果按照预设匹配规则插入预设面试信息存储模板。其中,预设面试信息存储模板中包括多个待填充槽位,例如该模板中包括待填充的具体姓名、具体学校、具体性别等,进而将实体结果与待填充槽位进行匹配,以将如“北京大学”填充至具体学校槽位中,进而形成 与目标面试者对应的面试信息页面。Specifically, after the interview sentence is input into the interview entity recognition model, the interview information words in the interview sentence are extracted and recognized, and the entity recognition result corresponding to each interview information word is obtained. The recognition result is inserted into the preset interview information storage template according to the preset matching rules. The preset interview information storage template includes multiple slots to be filled, for example, the template includes specific names, specific schools, specific genders, etc. to be filled, and then the entity results are matched with the slots to be filled, so as to "Peking University" is filled into the specific school slot, and then an interview information page corresponding to the target interviewee is formed.
应理解,上述实施例中各步骤的序号的大小并不意味着执行顺序的先后,各过程的执行顺序应以其功能和内在逻辑确定,而不应对本申请实施例的实施过程构成任何限定。It should be understood that the size of the sequence numbers of the steps in the above embodiments does not mean the sequence of execution, and the execution sequence of each process should be determined by its function and internal logic, and should not constitute any limitation to the implementation process of the embodiments of the present application.
在一实施例中,提供一种面试实体识别模型训练装置,该面试实体识别模型训练装置与上述实施例中面试实体识别模型训练方法一一对应。如图5所示,该面试实体识别模型训练装置包括样本数据获取模块10、标准标签预测模块20、辅助标签预测模块30、总损失值确定模块40和模型训练模块50。各功能模块详细说明如下:In one embodiment, an interview entity recognition model training device is provided, and the interview entity recognition model training device corresponds one-to-one with the interview entity recognition model training method in the above embodiment. As shown in FIG. 5 , the interview entity recognition model training device includes a sample data acquisition module 10 , a standard label prediction module 20 , an auxiliary label prediction module 30 , a total loss value determination module 40 and a model training module 50 . The detailed description of each functional module is as follows:
样本数据获取模块10,用于获取预设面试样本数据集;所述预设面试样本数据集中包含至少一个不具有面试标注标签的第一面试样本数据;A sample data acquisition module 10, configured to acquire a preset interview sample data set; the preset interview sample data set includes at least one first interview sample data without an interview label;
标准标签预测模块20,用于将所述第一面试样本数据输入至包含第一初始参数的预设识别模型中,通过所述预设识别模型中的直接预测模块对所述第一面试样本数据进行标准标签预测,得到标准标签分布以及与所述第一面试样本数据对应的面试编码向量;The standard label prediction module 20 is used to input the first interview sample data into a preset recognition model including the first initial parameter, and the first interview sample data is analyzed by the direct prediction module in the preset recognition model. Perform standard label prediction to obtain standard label distribution and interview coding vector corresponding to the first interview sample data;
辅助标签预测模块30,用于通过所述预设识别模型中的各辅助预测模块,根据所述编码向量对所述第一面试样本数据进行辅助标签预测,得到与各所述辅助预测模块输出的辅助标签分布;The auxiliary label prediction module 30 is configured to perform auxiliary label prediction on the first interview sample data according to the coding vector through each auxiliary prediction module in the preset recognition model, and obtain the output of each auxiliary prediction module. Auxiliary label distribution;
总损失值确定模块40,用于根据各所述辅助标签分布与所述标准标签分布确定所述预设识别模型的总损失值;A total loss value determination module 40, configured to determine the total loss value of the preset recognition model according to each of the auxiliary label distribution and the standard label distribution;
模型训练模块50,用于在所述总损失值未达到预设的收敛条件时,更新迭代所述预设识别模型的第一初始参数,直至所述总损失值达到所述预设的收敛条件时,将收敛之后的所述预设识别模型记录为面试实体识别模型。A model training module 50, configured to update and iterate the first initial parameter of the preset recognition model when the total loss value does not reach a preset convergence condition, until the total loss value reaches the preset convergence condition , the preset recognition model after convergence is recorded as the interview entity recognition model.
优选地,所述面试实体识别模型训练装置还包括:Preferably, the interview entity recognition model training device further includes:
直接标签预测模块,用于将所述第二面试样本数据输入至所述预设识别模型中,通过所述预设识别模型中包含第二初始参数的预设预测模块对所述第二面试样本数据进行直接标签预测,得到与所述第二面试样本数据对应的直接预测标签;The direct label prediction module is used to input the second interview sample data into the preset recognition model, and the second interview sample is analyzed by the preset prediction module including the second initial parameter in the preset recognition model. performing direct label prediction on the data to obtain a direct prediction label corresponding to the second interview sample data;
预测损失值确定模块,用于根据所述直接预测标签与所述面试标注标签确定所述预设编码器的预测损失值;a predicted loss value determination module, configured to determine the predicted loss value of the preset encoder according to the direct prediction label and the interview label;
参数更新模块,用于在所述预测损失值未达到预设的收敛条件时,更新迭代所述预设预设预测模块的第二初始参数,直至所述预测损失值达到所述预设的收敛条件时,将收敛之后的所述预设预测模块记录为所述直接预测模块。A parameter update module, configured to update and iterate the second initial parameter of the preset preset prediction module when the predicted loss value does not reach a preset convergence condition, until the predicted loss value reaches the preset convergence When conditions are met, the preset prediction module after convergence is recorded as the direct prediction module.
优选地,所述直接标签预测模块包括:Preferably, the direct label prediction module includes:
分词处理单元,用于对所述第二面试样本数据进行分词处理,得到与所述第二面试样本数据对应各标注样本字词;A word segmentation processing unit, configured to perform word segmentation processing on the second interview sample data to obtain each labeled sample word corresponding to the second interview sample data;
编码处理单元,用于通过所述预设预测模块中的编码器对各所述标注样本字词进行编码处理,得到第一正向编码向量以及第一反向编码向量;所述第一正向编码是指对各所述标注样本字词按照正向顺序进行编码得到;所述第一反向编码是指对各所述标注样本字词按照反向顺序进行编码得到;an encoding processing unit, configured to perform encoding processing on each of the labeled sample words by an encoder in the preset prediction module to obtain a first forward encoding vector and a first reverse encoding vector; the first forward encoding Encoding refers to encoding each of the labeled sample words in a forward order; the first reverse encoding refers to encoding each of the labeled sample words in a reverse order;
标签分类单元,用于根据所述第一正向编码向量以及所述第一反向编码向量,通过所述预设预测模块中的标注分类器对各所述标注样本字词进行标签分类,得到与各所述标注样本字词对应的直接预测标签。The label classification unit is configured to perform label classification on each of the labeled sample words by the labeling classifier in the preset prediction module according to the first forward coding vector and the first reverse coding vector, to obtain Directly predicted labels corresponding to each of the labeled sample words.
优选地,所述预测损失值确定模块包括:Preferably, the predicted loss value determination module includes:
样本实体标签获取单元,用于获取与各所述标注样本字词对应的所述样本实体标签;a sample entity label obtaining unit, configured to obtain the sample entity label corresponding to each of the labeled sample words;
标签损失值确定单元,用于根据与同一个标注样本字词对应的所述样本实体标签以及所述直接预测标签,确定与该标注样本字词对应的标签损失值;A label loss value determination unit, configured to determine a label loss value corresponding to the labeled sample word according to the sample entity label and the direct prediction label corresponding to the same labeled sample word;
预测损失值确定单元,用于根据与各所述标注样本字词对应的标签损失值,通过交差熵损失函数确定所述预测损失值。The predicted loss value determining unit is configured to determine the predicted loss value through the cross entropy loss function according to the label loss value corresponding to each of the labeled sample words.
优选地,如图6所示,所述辅助标签预测模块30包括:Preferably, as shown in FIG. 6 , the auxiliary label prediction module 30 includes:
编码向量获取单元301,用于获取所述编码向量中的至少两个第二正向编码向量以及至少两个第二反向编码向量;所述第二正向编码是指对各未标注样本字词按照正向顺序进行编码得到;所述第二反向编码是指对各所述未标注样本字词按照反向顺序进行编码得到;A coding vector obtaining unit 301, configured to obtain at least two second forward coding vectors and at least two second reverse coding vectors in the coding vectors; the second forward coding refers to The words are obtained by encoding in a forward order; the second reverse encoding refers to encoding each of the unlabeled sample words in a reverse order;
辅助标签分布确定单元302,用于根据各所述第二正向编码向量确定与各所述未标注样本字词对应的各正向辅助标签分布;同时,根据所述第二反向编码向量确定与各所述未标注样本字词对应的反向辅助标签分布。Auxiliary label distribution determining unit 302, configured to determine the distribution of forward auxiliary labels corresponding to each of the unlabeled sample words according to each of the second forward coding vectors; at the same time, determine according to the second reverse coding vector Reverse auxiliary label distribution corresponding to each of the unlabeled sample words.
在一实施例中,如图8所示,提供一种面试信息实体提取装置,包括:In one embodiment, as shown in FIG. 8, a device for extracting interview information entities is provided, including:
面试信息获取模块60,用于获取目标面试者的面试信息;所述面试信息中包含至少一个面试句子;一个所述面试句子中包含多个面试信息字词;The interview information obtaining module 60 is used to obtain the interview information of the target interviewee; the interview information includes at least one interview sentence; one of the interview sentences includes a plurality of interview information words;
实体提取识别模块70,用于将所述面试句子输入至面试实体识别模型中,对所述面试句子中的面试信息字词进行提取识别,得到与各所述面试信息字词对应的实体识别结果;所述面试实体识别模型是根据上述面试实体识别模型训练方法得到的;The entity extraction and recognition module 70 is used to input the interview sentence into the interview entity recognition model, extract and recognize the interview information words in the interview sentence, and obtain the entity recognition result corresponding to each interview information word ; Described interview entity recognition model is obtained according to above-mentioned interview entity recognition model training method;
信息存储模块80,用于将所述实体识别结果按照预设匹配规则插入预设面试信息存储模板。The information storage module 80 is configured to insert the entity recognition result into a preset interview information storage template according to preset matching rules.
关于面试实体识别模型训练装置以及面试信息实体提取装置的具体限定可以参见上文中对于面试实体识别模型训练方法以及面试信息实体提取方法的限定,在此不再赘述。上述面试实体识别模型训练装置以及面试信息实体提取装置中的各个模块可全部或部分通过软件、硬件及其组合来实现。上述各模块可以硬件形式内嵌于或独立于计算机设备中的处理器中,也可以以软件形式存储于计算机设备中的存储器中,以便于处理器调用执行以上各个模块对应的操作。For the specific limitations of the interview entity recognition model training device and the interview information entity extraction device, please refer to the limitations on the interview entity recognition model training method and the interview information entity extraction method above, which will not be repeated here. Each module in the above-mentioned interview entity recognition model training device and interview information entity extraction device can be implemented in whole or in part by software, hardware and combinations thereof. The above modules can be embedded in or independent of the processor in the computer device in the form of hardware, or stored in the memory in the computer device in the form of software, so that the processor can call and execute the operations corresponding to the above modules.
在一个实施例中,提供了一种计算机设备,该计算机设备可以是服务器,其内部结构图可以如图9所示。该计算机设备包括通过系统总线连接的处理器、存储器、网络接口和数据库。其中,该计算机设备的处理器用于提供计算和控制能力。该计算机设备的存储器包括可读存储介质、内存储器。该可读存储介质存储有操作系统、计算机可读指令和数据库。该内存储器为可读存储介质中的操作系统和计算机可读指令的运行提供环境。该计算机设备的数据库用于存储上述实施例中面试实体识别模型训练方法所使用到的数据、或者存储上述实施例中面试信息实体提取方法所使用到的数据。该计算机设备的网络接口用于与外部的终端通过网络连接通信。该计算机可读指令被处理器执行时以实现一种面试实体识别模型训练方法,或者该计算机可读指令被处理器执行时以实现一种面试信息实体提取方法。本实施例所提供的可读存储介质包括非易失性可读存储介质和易失性可读存储介质。In one embodiment, a computer device is provided, and the computer device may be a server, and its internal structure diagram may be as shown in FIG. 9 . The computer device includes a processor, memory, a network interface, and a database connected by a system bus. Among them, the processor of the computer device is used to provide computing and control capabilities. The memory of the computer device includes a readable storage medium, an internal memory. The readable storage medium stores an operating system, computer readable instructions and a database. The internal memory provides an environment for the execution of the operating system and computer-readable instructions in the readable storage medium. The database of the computer device is used to store the data used in the interview entity recognition model training method in the above-mentioned embodiment, or the data used in the interview information entity extraction method in the above-mentioned embodiment. The network interface of the computer device is used to communicate with an external terminal through a network connection. The computer-readable instructions, when executed by the processor, implement a method for training an interview entity recognition model, or the computer-readable instructions, when executed by the processor, implement a method for extracting interview information entities. The readable storage medium provided by this embodiment includes a non-volatile readable storage medium and a volatile readable storage medium.
在一实施例中,提供一种计算机设备,包括存储器、处理器以及存储在所述存储器中并可在所述处理器上运行的计算机可读指令,所述处理器执行所述计算机可读指令时实现如下步骤:In one embodiment, there is provided a computer apparatus comprising a memory, a processor, and computer readable instructions stored in the memory and executable on the processor, the processor executing the computer readable instructions When implementing the following steps:
获取预设面试样本数据集;所述预设面试样本数据集中包含至少一个不具有面试标注标签的第一面试样本数据;obtaining a preset interview sample data set; the preset interview sample data set includes at least one first interview sample data without an interview label;
将所述第一面试样本数据输入至包含第一初始参数的预设识别模型中,通过所述预设识别模型中的直接预测模块对所述第一面试样本数据进行标准标签预测,得到标准标签分布以及与所述第一面试样本数据对应的面试编码向量;Inputting the first interview sample data into a preset recognition model including the first initial parameter, and performing standard label prediction on the first interview sample data through the direct prediction module in the preset recognition model to obtain a standard label distribution and an interview coding vector corresponding to the first interview sample data;
通过所述预设识别模型中的各辅助预测模块,根据所述面试编码向量对所述第一面试样本数据进行辅助标签预测,得到与各所述辅助预测模块输出的辅助标签分布;Through each auxiliary prediction module in the preset recognition model, carry out auxiliary label prediction on the first interview sample data according to the interview coding vector, and obtain the auxiliary label distribution output by each of the auxiliary prediction modules;
根据各所述辅助标签分布与所述标准标签分布确定所述预设识别模型的总损失值;在所述总损失值未达到预设的收敛条件时,更新迭代所述预设识别模型的第一初始参数, 直至所述总损失值达到所述预设的收敛条件时,将收敛之后的所述预设识别模型记录为面试实体识别模型。Determine the total loss value of the preset recognition model according to the distribution of each auxiliary label and the standard label distribution; when the total loss value does not reach the preset convergence condition, update the first step of iterating the preset recognition model. an initial parameter, until the total loss value reaches the preset convergence condition, the preset recognition model after convergence is recorded as the interview entity recognition model.
在一实施例中,提供一种计算机设备,包括存储器、处理器以及存储在所述存储器中并可在所述处理器上运行的计算机可读指令,所述处理器执行所述计算机可读指令时实现如下步骤:In one embodiment, there is provided a computer apparatus comprising a memory, a processor, and computer readable instructions stored in the memory and executable on the processor, the processor executing the computer readable instructions When implementing the following steps:
获取目标面试者的面试信息;所述面试信息中包含至少一个面试句子;一个所述面试句子中包含多个面试信息字词;Obtain the interview information of the target interviewee; the interview information includes at least one interview sentence; one of the interview sentences includes a plurality of interview information words;
将所述面试句子输入至面试实体识别模型中,对所述面试句子中的面试信息字词进行提取识别,得到与各所述面试信息字词对应的实体识别结果;所述面试实体识别模型是根据上述面试实体识别模型训练方法得到的;The interview sentence is input into the interview entity recognition model, and the interview information words in the interview sentence are extracted and recognized to obtain entity recognition results corresponding to the interview information words; the interview entity recognition model is Obtained according to the above interview entity recognition model training method;
将所述实体识别结果按照预设匹配规则插入预设面试信息存储模板。Insert the entity recognition result into a preset interview information storage template according to preset matching rules.
在一实施例中,提供一个或多个存储有计算机可读指令的可读存储介质,所述计算机可读指令被一个或多个处理器执行时,使得所述一个或多个处理器执行如下步骤:In one embodiment, one or more readable storage media are provided storing computer-readable instructions that, when executed by one or more processors, cause the one or more processors to perform the following: step:
获取预设面试样本数据集;所述预设面试样本数据集中包含至少一个不具有面试标注标签的第一面试样本数据;obtaining a preset interview sample data set; the preset interview sample data set includes at least one first interview sample data without an interview label;
将所述第一面试样本数据输入至包含第一初始参数的预设识别模型中,通过所述预设识别模型中的直接预测模块对所述第一面试样本数据进行标准标签预测,得到标准标签分布以及与所述第一面试样本数据对应的面试编码向量;Inputting the first interview sample data into a preset recognition model including the first initial parameter, and performing standard label prediction on the first interview sample data through the direct prediction module in the preset recognition model to obtain a standard label distribution and an interview coding vector corresponding to the first interview sample data;
通过所述预设识别模型中的各辅助预测模块,根据所述面试编码向量对所述第一面试样本数据进行辅助标签预测,得到与各所述辅助预测模块输出的辅助标签分布;Through each auxiliary prediction module in the preset recognition model, carry out auxiliary label prediction on the first interview sample data according to the interview coding vector, and obtain the auxiliary label distribution output by each of the auxiliary prediction modules;
根据各所述辅助标签分布与所述标准标签分布确定所述预设识别模型的总损失值;Determine the total loss value of the preset recognition model according to each of the auxiliary label distribution and the standard label distribution;
在所述总损失值未达到预设的收敛条件时,更新迭代所述预设识别模型的第一初始参数,直至所述总损失值达到所述预设的收敛条件时,将收敛之后的所述预设识别模型记录为面试实体识别模型。When the total loss value does not reach the preset convergence condition, update and iterate the first initial parameter of the preset recognition model, until the total loss value reaches the preset convergence condition, all the The aforementioned preset recognition model is recorded as the interview entity recognition model.
在一实施例中,提供一个或多个存储有计算机可读指令的可读存储介质,所述计算机可读指令被一个或多个处理器执行时,使得所述一个或多个处理器执行如下步骤:In one embodiment, one or more readable storage media are provided storing computer-readable instructions that, when executed by one or more processors, cause the one or more processors to perform the following: step:
获取目标面试者的面试信息;所述面试信息中包含至少一个面试句子;一个所述面试句子中包含多个面试信息字词;Obtain the interview information of the target interviewee; the interview information includes at least one interview sentence; one of the interview sentences includes a plurality of interview information words;
将所述面试句子输入至面试实体识别模型中,对所述面试句子中的面试信息字词进行提取识别,得到与各所述面试信息字词对应的实体识别结果;所述面试实体识别模型是根据上述面试实体识别模型训练方法得到的;The interview sentence is input into the interview entity recognition model, and the interview information words in the interview sentence are extracted and recognized to obtain entity recognition results corresponding to the interview information words; the interview entity recognition model is Obtained according to the above interview entity recognition model training method;
将所述实体识别结果按照预设匹配规则插入预设面试信息存储模板。Insert the entity recognition result into a preset interview information storage template according to preset matching rules.
本领域普通技术人员可以理解实现上述实施例方法中的全部或部分流程,是可以通过计算机可读指令来指令相关的硬件来完成,所述的计算机可读指令可存储于一非易失性计算机可读取存储介质或者易失性计算机可读取存储介质中,该计算机可读指令在执行时,可包括如上述各方法的实施例的流程。其中,本申请所提供的各实施例中所使用的对存储器、存储、数据库或其它介质的任何引用,均可包括非易失性和/或易失性存储器。非易失性存储器可包括只读存储器(ROM)、可编程ROM(PROM)、电可编程ROM(EPROM)、电可擦除可编程ROM(EEPROM)或闪存。易失性存储器可包括随机存取存储器(RAM)或者外部高速缓冲存储器。作为说明而非局限,RAM以多种形式可得,诸如静态RAM(SRAM)、动态RAM(DRAM)、同步DRAM(SDRAM)、双数据率SDRAM(DDRSDRAM)、增强型SDRAM(ESDRAM)、同步链路(Synchlink)DRAM(SLDRAM)、存储器总线(Rambus)直接RAM(RDRAM)、直接存储器总线动态RAM(DRDRAM)、以及存储器总线动态RAM(RDRAM)等。Those of ordinary skill in the art can understand that all or part of the processes in the methods of the above embodiments can be implemented by instructing relevant hardware through computer-readable instructions, and the computer-readable instructions can be stored in a non-volatile computer. In a readable storage medium or a volatile computer-readable storage medium, the computer-readable instructions, when executed, may include the processes of the foregoing method embodiments. Wherein, any reference to memory, storage, database or other medium used in the various embodiments provided in this application may include non-volatile and/or volatile memory. Nonvolatile memory may include read only memory (ROM), programmable ROM (PROM), electrically programmable ROM (EPROM), electrically erasable programmable ROM (EEPROM), or flash memory. Volatile memory may include random access memory (RAM) or external cache memory. By way of illustration and not limitation, RAM is available in various forms such as static RAM (SRAM), dynamic RAM (DRAM), synchronous DRAM (SDRAM), double data rate SDRAM (DDRSDRAM), enhanced SDRAM (ESDRAM), synchronous chain Road (Synchlink) DRAM (SLDRAM), memory bus (Rambus) direct RAM (RDRAM), direct memory bus dynamic RAM (DRDRAM), and memory bus dynamic RAM (RDRAM), etc.
所属领域的技术人员可以清楚地了解到,为了描述的方便和简洁,仅以上述各功能单 元、模块的划分进行举例说明,实际应用中,可以根据需要而将上述功能分配由不同的功能单元、模块完成,即将所述装置的内部结构划分成不同的功能单元或模块,以完成以上描述的全部或者部分功能。Those skilled in the art can clearly understand that, for the convenience and simplicity of description, only the division of the above-mentioned functional units and modules is used as an example. Module completion, that is, dividing the internal structure of the device into different functional units or modules to complete all or part of the functions described above.
以上所述实施例仅用以说明本申请的技术方案,而非对其限制;尽管参照前述实施例对本申请进行了详细的说明,本领域的普通技术人员应当理解:其依然可以对前述各实施例所记载的技术方案进行修改,或者对其中部分技术特征进行等同替换;而这些修改或者替换,并不使相应技术方案的本质脱离本申请各实施例技术方案的精神和范围,均应包含在本申请的保护范围之内。The above-mentioned embodiments are only used to illustrate the technical solutions of the present application, but not to limit them; although the present application has been described in detail with reference to the above-mentioned embodiments, those of ordinary skill in the art should understand that: it can still be used for the above-mentioned implementations. The technical solutions described in the examples are modified, or some technical features thereof are equivalently replaced; and these modifications or replacements do not make the essence of the corresponding technical solutions deviate from the spirit and scope of the technical solutions in the embodiments of the application, and should be included in the within the scope of protection of this application.

Claims (20)

  1. 一种面试实体识别模型训练方法,其中,包括:An interview entity recognition model training method, which includes:
    获取预设面试样本数据集;所述预设面试样本数据集中包含至少一个不具有面试标注标签的第一面试样本数据;obtaining a preset interview sample data set; the preset interview sample data set includes at least one first interview sample data without an interview label;
    将所述第一面试样本数据输入至包含第一初始参数的预设识别模型中,通过所述预设识别模型中的直接预测模块对所述第一面试样本数据进行标准标签预测,得到标准标签分布以及与所述第一面试样本数据对应的面试编码向量;Inputting the first interview sample data into a preset recognition model including the first initial parameter, and performing standard label prediction on the first interview sample data through the direct prediction module in the preset recognition model to obtain a standard label distribution and an interview coding vector corresponding to the first interview sample data;
    通过所述预设识别模型中的各辅助预测模块,根据所述面试编码向量对所述第一面试样本数据进行辅助标签预测,得到与各所述辅助预测模块输出的辅助标签分布;Through each auxiliary prediction module in the preset recognition model, carry out auxiliary label prediction on the first interview sample data according to the interview coding vector, and obtain the auxiliary label distribution output by each of the auxiliary prediction modules;
    根据各所述辅助标签分布与所述标准标签分布确定所述预设识别模型的总损失值;Determine the total loss value of the preset recognition model according to each of the auxiliary label distribution and the standard label distribution;
    在所述总损失值未达到预设的收敛条件时,更新迭代所述预设识别模型的第一初始参数,直至所述总损失值达到所述预设的收敛条件时,将收敛之后的所述预设识别模型记录为面试实体识别模型。When the total loss value does not reach the preset convergence condition, update and iterate the first initial parameter of the preset recognition model, until the total loss value reaches the preset convergence condition, all the The aforementioned preset recognition model is recorded as the interview entity recognition model.
  2. 如权利要求1所述的面试实体识别模型训练方法,其中,所述预设面试样本数据集中还包含至少一个具有所述面试标注标签的第二面试样本数据;所述通过所述预设识别模型中的预设编码器对所述第一面试样本数据进行标准标签预测之前,包括:The method for training an interview entity recognition model according to claim 1, wherein the preset interview sample data set further includes at least one second interview sample data with the interview label; Before performing standard label prediction on the first interview sample data, the preset encoder in , includes:
    将所述第二面试样本数据输入至所述预设识别模型中,通过所述预设识别模型中包含第二初始参数的预设预测模块对所述第二面试样本数据进行直接标签预测,得到与所述第二面试样本数据对应的直接预测标签;Input the second interview sample data into the preset recognition model, and perform direct label prediction on the second interview sample data through the preset prediction module that includes the second initial parameter in the preset recognition model, to obtain the direct prediction label corresponding to the second interview sample data;
    根据所述直接预测标签与所述面试标注标签确定所述预设编码器的预测损失值;Determine the prediction loss value of the preset encoder according to the direct prediction label and the interview label;
    在所述预测损失值未达到预设的收敛条件时,更新迭代所述预设预设预测模块的第二初始参数,直至所述预测损失值达到所述预设的收敛条件时,将收敛之后的所述预设预测模块记录为所述直接预测模块。When the predicted loss value does not reach the preset convergence condition, update and iterate the second initial parameter of the preset preset prediction module until the predicted loss value reaches the preset convergence condition, after the convergence The preset prediction module is recorded as the direct prediction module.
  3. 如权利要求2所述的面试实体识别模型训练方法,其中,所述通过所述预设识别模型中包含第二初始参数的预设预测模块对所述第二面试样本数据进行直接标签预测,得到与所述第二面试样本数据对应的直接预测标签,包括:The method for training an interview entity recognition model according to claim 2, wherein the direct label prediction is performed on the second interview sample data by using a preset prediction module including a second initial parameter in the preset recognition model, to obtain The direct prediction label corresponding to the second interview sample data, including:
    对所述第二面试样本数据进行分词处理,得到与所述第二面试样本数据对应各标注样本字词;Perform word segmentation processing on the second interview sample data to obtain each marked sample word corresponding to the second interview sample data;
    通过所述预设预测模块中的编码器对各所述标注样本字词进行编码处理,得到第一正向编码向量以及第一反向编码向量;所述第一正向编码是指对各所述标注样本字词按照正向顺序进行编码得到;所述第一反向编码是指对各所述标注样本字词按照反向顺序进行编码得到;The encoder in the preset prediction module performs encoding processing on each of the labeled sample words to obtain a first forward encoding vector and a first reverse encoding vector; the first forward encoding refers to The labeled sample words are obtained by encoding in a forward order; the first reverse encoding refers to encoding each of the labeled sample words in a reverse order;
    根据所述第一正向编码向量以及所述第一反向编码向量,通过所述预设预测模块中的标注分类器对各所述标注样本字词进行标签分类,得到与各所述标注样本字词对应的直接预测标签。According to the first forward coding vector and the first reverse coding vector, the labeling classifier in the preset prediction module is used to label each of the labelled sample words, and obtain the same label as each labelled sample. The direct prediction label corresponding to the term.
  4. 如权利要求2所述的面试实体识别模型训练方法,其中,所述面试标注标签中包含多个样本实体标签;所述根据所述直接预测标签与所述面试标注标签确定所述预设编码器的预测损失值,包括:The method for training an interview entity recognition model according to claim 2, wherein the interview label includes a plurality of sample entity labels; the preset encoder is determined according to the direct prediction label and the interview label The predicted loss value of , including:
    获取与各所述标注样本字词对应的所述样本实体标签;obtaining the sample entity labels corresponding to each of the labeled sample words;
    根据与同一个标注样本字词对应的所述样本实体标签以及所述直接预测标签,确定与该标注样本字词对应的标签损失值;Determine the label loss value corresponding to the labeled sample word according to the sample entity label and the directly predicted label corresponding to the same labeled sample word;
    根据与各所述标注样本字词对应的标签损失值,通过交差熵损失函数确定所述预测损失值。According to the label loss value corresponding to each of the labeled sample words, the predicted loss value is determined through a cross entropy loss function.
  5. 如权利要求1所述的面试实体识别模型训练方法,其中,所述通过所述预设识别模型中的各辅助预测模块,根据所述编码向量对所述第一面试样本数据进行辅助标签预测,得到与各所述辅助预测模块输出的辅助标签分布,包括:The method for training an interview entity recognition model according to claim 1, wherein the auxiliary label prediction is performed on the first interview sample data according to the coding vector through each auxiliary prediction module in the preset recognition model, Obtain the auxiliary label distribution output by each of the auxiliary prediction modules, including:
    获取所述面试编码向量中的至少两个第二正向编码向量以及至少两个第二反向编码向量;所述第二正向编码是指对各未标注样本字词按照正向顺序进行编码得到;所述第二反向编码是指对各所述未标注样本字词按照反向顺序进行编码得到;Obtain at least two second forward encoding vectors and at least two second reverse encoding vectors in the interview encoding vectors; the second forward encoding refers to encoding each unlabeled sample word in a forward order Obtained; the second reverse encoding refers to encoding each of the unlabeled sample words in reverse order to obtain;
    根据各所述第二正向编码向量确定与各所述未标注样本字词对应的各正向辅助标签分布;同时,根据所述第二反向编码向量确定与各所述未标注样本字词对应的反向辅助标签分布。According to each of the second forward coding vectors, determine the distribution of each forward auxiliary label corresponding to each of the unlabeled sample words; at the same time, determine the distribution of each of the unlabeled sample words according to the second reverse coding vector. The corresponding reverse auxiliary label distribution.
  6. 一种面试信息实体提取方法,其中,包括:A method for extracting interview information entities, comprising:
    获取目标面试者的面试信息;所述面试信息中包含至少一个面试句子;一个所述面试句子中包含多个面试信息字词;Obtain the interview information of the target interviewee; the interview information includes at least one interview sentence; one of the interview sentences includes a plurality of interview information words;
    将所述面试句子输入至面试实体识别模型中,对所述面试句子中的面试信息字词进行提取识别,得到与各所述面试信息字词对应的实体识别结果;所述面试实体识别模型是根据如权利要求1至5任一项所述面试实体识别模型训练方法得到的;The interview sentence is input into the interview entity recognition model, and the interview information words in the interview sentence are extracted and recognized to obtain entity recognition results corresponding to the interview information words; the interview entity recognition model is Obtained according to the interview entity recognition model training method as described in any one of claims 1 to 5;
    将所述实体识别结果按照预设匹配规则插入预设面试信息存储模板。Insert the entity recognition result into a preset interview information storage template according to preset matching rules.
  7. 一种面试实体识别模型训练装置,其中,包括:An interview entity recognition model training device, comprising:
    样本数据获取模块,用于获取预设面试样本数据集;所述预设面试样本数据集中包含至少一个不具有面试标注标签的第一面试样本数据;a sample data acquisition module, configured to acquire a preset interview sample data set; the preset interview sample data set includes at least one first interview sample data without an interview label;
    标准标签预测模块,用于将所述第一面试样本数据输入至包含第一初始参数的预设识别模型中,通过所述预设识别模型中的直接预测模块对所述第一面试样本数据进行标准标签预测,得到标准标签分布以及与所述第一面试样本数据对应的面试编码向量;The standard label prediction module is used to input the first interview sample data into a preset recognition model including the first initial parameter, and perform the first interview sample data through the direct prediction module in the preset recognition model. Standard label prediction, to obtain standard label distribution and interview coding vector corresponding to the first interview sample data;
    辅助标签预测模块,用于通过所述预设识别模型中的各辅助预测模块,根据所述编码向量对所述第一面试样本数据进行辅助标签预测,得到与各所述辅助预测模块输出的辅助标签分布;The auxiliary label prediction module is used to perform auxiliary label prediction on the first interview sample data according to the coding vector through each auxiliary prediction module in the preset recognition model, and obtain the auxiliary label output with each of the auxiliary prediction modules. label distribution;
    总损失值确定模块,用于根据各所述辅助标签分布与所述标准标签分布确定所述预设识别模型的总损失值;a total loss value determination module, configured to determine the total loss value of the preset recognition model according to each of the auxiliary label distributions and the standard label distribution;
    模型训练模块,用于在所述总损失值未达到预设的收敛条件时,更新迭代所述预设识别模型的第一初始参数,直至所述总损失值达到所述预设的收敛条件时,将收敛之后的所述预设识别模型记录为面试实体识别模型。A model training module, configured to update and iterate the first initial parameter of the preset recognition model when the total loss value does not reach the preset convergence condition, until the total loss value reaches the preset convergence condition , and record the preset recognition model after convergence as an interview entity recognition model.
  8. 一种面试信息实体提取装置,其中,包括:A device for extracting interview information entities, comprising:
    面试信息获取模块,用于获取目标面试者的面试信息;所述面试信息中包含至少一个面试句子;一个所述面试句子中包含多个面试信息字词;an interview information obtaining module, used for obtaining interview information of a target interviewee; the interview information includes at least one interview sentence; one interview sentence includes a plurality of interview information words;
    实体提取识别模块,用于将所述面试句子输入至面试实体识别模型中,对所述面试句子中的面试信息字词进行提取识别,得到与各所述面试信息字词对应的实体识别结果;所述面试实体识别模型是根据如权利要求1至5任一项所述面试实体识别模型训练方法得到的;An entity extraction and recognition module is used to input the interview sentence into the interview entity recognition model, extract and recognize the interview information words in the interview sentence, and obtain entity recognition results corresponding to each of the interview information words; The interview entity recognition model is obtained according to the interview entity recognition model training method as described in any one of claims 1 to 5;
    信息存储模块,用于将所述实体识别结果按照预设匹配规则插入预设面试信息存储模板。An information storage module, configured to insert the entity recognition result into a preset interview information storage template according to preset matching rules.
  9. 一种计算机设备,包括存储器、处理器以及存储在所述存储器中并可在所述处理器上运行的计算机可读指令,其中,所述处理器执行所述计算机可读指令时实现如下步骤:A computer device comprising a memory, a processor, and computer-readable instructions stored in the memory and executable on the processor, wherein the processor implements the following steps when executing the computer-readable instructions:
    获取预设面试样本数据集;所述预设面试样本数据集中包含至少一个不具有面试标注标签的第一面试样本数据;obtaining a preset interview sample data set; the preset interview sample data set includes at least one first interview sample data without an interview label;
    将所述第一面试样本数据输入至包含第一初始参数的预设识别模型中,通过所述预设识别模型中的直接预测模块对所述第一面试样本数据进行标准标签预测,得到标准标签分 布以及与所述第一面试样本数据对应的面试编码向量;Inputting the first interview sample data into a preset recognition model including the first initial parameter, and performing standard label prediction on the first interview sample data through the direct prediction module in the preset recognition model to obtain a standard label distribution and an interview coding vector corresponding to the first interview sample data;
    通过所述预设识别模型中的各辅助预测模块,根据所述面试编码向量对所述第一面试样本数据进行辅助标签预测,得到与各所述辅助预测模块输出的辅助标签分布;Through each auxiliary prediction module in the preset recognition model, carry out auxiliary label prediction on the first interview sample data according to the interview coding vector, and obtain the auxiliary label distribution output by each of the auxiliary prediction modules;
    根据各所述辅助标签分布与所述标准标签分布确定所述预设识别模型的总损失值;在所述总损失值未达到预设的收敛条件时,更新迭代所述预设识别模型的第一初始参数,直至所述总损失值达到所述预设的收敛条件时,将收敛之后的所述预设识别模型记录为面试实体识别模型。Determine the total loss value of the preset recognition model according to the distribution of each auxiliary label and the standard label distribution; when the total loss value does not reach the preset convergence condition, update the first step of iterating the preset recognition model. An initial parameter, until the total loss value reaches the preset convergence condition, the preset recognition model after convergence is recorded as the interview entity recognition model.
  10. 如权利要求9所述的计算机设备,其中,所述预设面试样本数据集中还包含至少一个具有所述面试标注标签的第二面试样本数据;所述通过所述预设识别模型中的预设编码器对所述第一面试样本数据进行标准标签预测之前,所述处理器执行所述计算机可读指令时还实现如下步骤:The computer device according to claim 9, wherein the preset interview sample data set further includes at least one second interview sample data with the interview label; Before the encoder performs standard label prediction on the first interview sample data, the processor further implements the following steps when executing the computer-readable instructions:
    将所述第二面试样本数据输入至所述预设识别模型中,通过所述预设识别模型中包含第二初始参数的预设预测模块对所述第二面试样本数据进行直接标签预测,得到与所述第二面试样本数据对应的直接预测标签;Input the second interview sample data into the preset recognition model, and perform direct label prediction on the second interview sample data through the preset prediction module that includes the second initial parameter in the preset recognition model, to obtain the direct prediction label corresponding to the second interview sample data;
    根据所述直接预测标签与所述面试标注标签确定所述预设编码器的预测损失值;Determine the prediction loss value of the preset encoder according to the direct prediction label and the interview label;
    在所述预测损失值未达到预设的收敛条件时,更新迭代所述预设预设预测模块的第二初始参数,直至所述预测损失值达到所述预设的收敛条件时,将收敛之后的所述预设预测模块记录为所述直接预测模块。When the predicted loss value does not reach the preset convergence condition, update and iterate the second initial parameter of the preset preset prediction module until the predicted loss value reaches the preset convergence condition, after the convergence The preset prediction module is recorded as the direct prediction module.
  11. 如权利要求10所述的计算机设备,其中,所述通过所述预设识别模型中包含第二初始参数的预设预测模块对所述第二面试样本数据进行直接标签预测,得到与所述第二面试样本数据对应的直接预测标签,包括:The computer device according to claim 10, wherein the direct label prediction is performed on the second interview sample data through a preset prediction module including a second initial parameter in the preset recognition model, and a result that is the same as the first one is obtained. The direct prediction labels corresponding to the second interview sample data, including:
    对所述第二面试样本数据进行分词处理,得到与所述第二面试样本数据对应各标注样本字词;Perform word segmentation processing on the second interview sample data to obtain each marked sample word corresponding to the second interview sample data;
    通过所述预设预测模块中的编码器对各所述标注样本字词进行编码处理,得到第一正向编码向量以及第一反向编码向量;所述第一正向编码是指对各所述标注样本字词按照正向顺序进行编码得到;所述第一反向编码是指对各所述标注样本字词按照反向顺序进行编码得到;The encoder in the preset prediction module performs encoding processing on each of the labeled sample words to obtain a first forward encoding vector and a first reverse encoding vector; the first forward encoding refers to The labeled sample words are obtained by encoding in a forward order; the first reverse encoding refers to encoding each of the labeled sample words in a reverse order;
    根据所述第一正向编码向量以及所述第一反向编码向量,通过所述预设预测模块中的标注分类器对各所述标注样本字词进行标签分类,得到与各所述标注样本字词对应的直接预测标签。According to the first forward coding vector and the first reverse coding vector, the labeling classifier in the preset prediction module is used to label each of the labelled sample words, and obtain the same label as each labelled sample. The direct prediction label corresponding to the term.
  12. 如权利要求10所述的计算机设备,其中,所述面试标注标签中包含多个样本实体标签;所述根据所述直接预测标签与所述面试标注标签确定所述预设编码器的预测损失值,包括:The computer device according to claim 10, wherein the interview annotation label includes a plurality of sample entity labels; the prediction loss value of the preset encoder is determined according to the direct prediction label and the interview annotation label ,include:
    获取与各所述标注样本字词对应的所述样本实体标签;obtaining the sample entity labels corresponding to each of the labeled sample words;
    根据与同一个标注样本字词对应的所述样本实体标签以及所述直接预测标签,确定与该标注样本字词对应的标签损失值;Determine the label loss value corresponding to the labeled sample word according to the sample entity label and the directly predicted label corresponding to the same labeled sample word;
    根据与各所述标注样本字词对应的标签损失值,通过交差熵损失函数确定所述预测损失值。According to the label loss value corresponding to each of the labeled sample words, the predicted loss value is determined through a cross entropy loss function.
  13. 如权利要求9所述的计算机设备,其中,所述通过所述预设识别模型中的各辅助预测模块,根据所述编码向量对所述第一面试样本数据进行辅助标签预测,得到与各所述辅助预测模块输出的辅助标签分布,包括:The computer device according to claim 9, wherein the auxiliary label prediction is performed on the first interview sample data according to the coding vector through each auxiliary prediction module in the preset recognition model, and the corresponding prediction is obtained by each auxiliary prediction module. The auxiliary label distribution output by the auxiliary prediction module, including:
    获取所述面试编码向量中的至少两个第二正向编码向量以及至少两个第二反向编码向量;所述第二正向编码是指对各未标注样本字词按照正向顺序进行编码得到;所述第二反向编码是指对各所述未标注样本字词按照反向顺序进行编码得到;Obtain at least two second forward encoding vectors and at least two second reverse encoding vectors in the interview encoding vectors; the second forward encoding refers to encoding each unlabeled sample word in a forward order Obtained; the second reverse encoding refers to encoding each of the unlabeled sample words in reverse order to obtain;
    根据各所述第二正向编码向量确定与各所述未标注样本字词对应的各正向辅助标签分布;同时,根据所述第二反向编码向量确定与各所述未标注样本字词对应的反向辅助标签分布。According to each of the second forward coding vectors, determine the distribution of each forward auxiliary label corresponding to each of the unlabeled sample words; at the same time, determine the distribution of each of the unlabeled sample words according to the second reverse coding vector. The corresponding reverse auxiliary label distribution.
  14. 一种计算机设备,包括存储器、处理器以及存储在所述存储器中并可在所述处理器上运行的计算机可读指令,其中,所述处理器执行所述计算机可读指令时实现如下步骤:A computer device comprising a memory, a processor, and computer-readable instructions stored in the memory and executable on the processor, wherein the processor implements the following steps when executing the computer-readable instructions:
    获取目标面试者的面试信息;所述面试信息中包含至少一个面试句子;一个所述面试句子中包含多个面试信息字词;Obtain interview information of the target interviewee; the interview information includes at least one interview sentence; one interview sentence includes multiple interview information words;
    将所述面试句子输入至面试实体识别模型中,对所述面试句子中的面试信息字词进行提取识别,得到与各所述面试信息字词对应的实体识别结果;所述面试实体识别模型是根据如权利要求1至5任一项所述面试实体识别模型训练方法得到的;The interview sentence is input into the interview entity recognition model, and the interview information words in the interview sentence are extracted and recognized to obtain entity recognition results corresponding to the interview information words; the interview entity recognition model is Obtained according to the interview entity recognition model training method as described in any one of claims 1 to 5;
    将所述实体识别结果按照预设匹配规则插入预设面试信息存储模板。Insert the entity recognition result into a preset interview information storage template according to preset matching rules.
  15. 一个或多个存储有计算机可读指令的可读存储介质,其中,所述计算机可读指令被一个或多个处理器执行时,使得所述一个或多个处理器执行如下步骤:One or more readable storage media storing computer-readable instructions, wherein the computer-readable instructions, when executed by one or more processors, cause the one or more processors to perform the following steps:
    获取预设面试样本数据集;所述预设面试样本数据集中包含至少一个不具有面试标注标签的第一面试样本数据;obtaining a preset interview sample data set; the preset interview sample data set includes at least one first interview sample data without an interview label;
    将所述第一面试样本数据输入至包含第一初始参数的预设识别模型中,通过所述预设识别模型中的直接预测模块对所述第一面试样本数据进行标准标签预测,得到标准标签分布以及与所述第一面试样本数据对应的面试编码向量;Inputting the first interview sample data into a preset recognition model including the first initial parameter, and performing standard label prediction on the first interview sample data through the direct prediction module in the preset recognition model to obtain a standard label distribution and an interview coding vector corresponding to the first interview sample data;
    通过所述预设识别模型中的各辅助预测模块,根据所述面试编码向量对所述第一面试样本数据进行辅助标签预测,得到与各所述辅助预测模块输出的辅助标签分布;Through each auxiliary prediction module in the preset recognition model, carry out auxiliary label prediction on the first interview sample data according to the interview coding vector, and obtain the auxiliary label distribution output by each of the auxiliary prediction modules;
    根据各所述辅助标签分布与所述标准标签分布确定所述预设识别模型的总损失值;Determine the total loss value of the preset recognition model according to each of the auxiliary label distribution and the standard label distribution;
    在所述总损失值未达到预设的收敛条件时,更新迭代所述预设识别模型的第一初始参数,直至所述总损失值达到所述预设的收敛条件时,将收敛之后的所述预设识别模型记录为面试实体识别模型。When the total loss value does not reach the preset convergence condition, update and iterate the first initial parameter of the preset recognition model, until the total loss value reaches the preset convergence condition, all the The aforementioned preset recognition model is recorded as the interview entity recognition model.
  16. 如权利要求15所述的可读存储介质,其中,所述预设面试样本数据集中还包含至少一个具有所述面试标注标签的第二面试样本数据;所述通过所述预设识别模型中的预设编码器对所述第一面试样本数据进行标准标签预测之前,所述计算机可读指令被一个或多个处理器执行时,使得所述一个或多个处理器还执行如下步骤:The readable storage medium according to claim 15, wherein the preset interview sample data set further includes at least one second interview sample data with the interview label; Before the preset encoder performs standard label prediction on the first interview sample data, when the computer-readable instructions are executed by one or more processors, the one or more processors further perform the following steps:
    将所述第二面试样本数据输入至所述预设识别模型中,通过所述预设识别模型中包含第二初始参数的预设预测模块对所述第二面试样本数据进行直接标签预测,得到与所述第二面试样本数据对应的直接预测标签;Input the second interview sample data into the preset recognition model, and perform direct label prediction on the second interview sample data through the preset prediction module that includes the second initial parameter in the preset recognition model, to obtain the direct prediction label corresponding to the second interview sample data;
    根据所述直接预测标签与所述面试标注标签确定所述预设编码器的预测损失值;Determine the prediction loss value of the preset encoder according to the direct prediction label and the interview label;
    在所述预测损失值未达到预设的收敛条件时,更新迭代所述预设预设预测模块的第二初始参数,直至所述预测损失值达到所述预设的收敛条件时,将收敛之后的所述预设预测模块记录为所述直接预测模块。When the predicted loss value does not reach the preset convergence condition, update and iterate the second initial parameter of the preset preset prediction module until the predicted loss value reaches the preset convergence condition, after the convergence The preset prediction module is recorded as the direct prediction module.
  17. 如权利要求16所述的可读存储介质,其中,所述通过所述预设识别模型中包含第二初始参数的预设预测模块对所述第二面试样本数据进行直接标签预测,得到与所述第二面试样本数据对应的直接预测标签,包括:The readable storage medium according to claim 16, wherein the direct label prediction is performed on the second interview sample data by a preset prediction module including the second initial parameter in the preset recognition model, and the result is obtained with the The direct prediction labels corresponding to the second interview sample data, including:
    对所述第二面试样本数据进行分词处理,得到与所述第二面试样本数据对应各标注样本字词;Perform word segmentation processing on the second interview sample data to obtain each marked sample word corresponding to the second interview sample data;
    通过所述预设预测模块中的编码器对各所述标注样本字词进行编码处理,得到第一正向编码向量以及第一反向编码向量;所述第一正向编码是指对各所述标注样本字词按照正向顺序进行编码得到;所述第一反向编码是指对各所述标注样本字词按照反向顺序进行编码得到;The encoder in the preset prediction module performs encoding processing on each of the labeled sample words to obtain a first forward encoding vector and a first reverse encoding vector; the first forward encoding refers to The labeled sample words are obtained by encoding in a forward order; the first reverse encoding refers to encoding each of the labeled sample words in a reverse order;
    根据所述第一正向编码向量以及所述第一反向编码向量,通过所述预设预测模块中的标注分类器对各所述标注样本字词进行标签分类,得到与各所述标注样本字词对应的直接预测标签。According to the first forward coding vector and the first reverse coding vector, the labeling classifier in the preset prediction module is used to label each of the labelled sample words, and obtain the same label as each labelled sample. The direct prediction label corresponding to the term.
  18. 如权利要求16所述的可读存储介质,其中,所述面试标注标签中包含多个样本实体标签;所述根据所述直接预测标签与所述面试标注标签确定所述预设编码器的预测损失值,包括:The readable storage medium according to claim 16, wherein the interview annotation label includes a plurality of sample entity labels; the prediction of the preset encoder is determined according to the direct prediction label and the interview annotation label Loss values, including:
    获取与各所述标注样本字词对应的所述样本实体标签;obtaining the sample entity labels corresponding to each of the labeled sample words;
    根据与同一个标注样本字词对应的所述样本实体标签以及所述直接预测标签,确定与该标注样本字词对应的标签损失值;Determine the label loss value corresponding to the labeled sample word according to the sample entity label and the directly predicted label corresponding to the same labeled sample word;
    根据与各所述标注样本字词对应的标签损失值,通过交差熵损失函数确定所述预测损失值。According to the label loss value corresponding to each of the labeled sample words, the predicted loss value is determined through a cross entropy loss function.
  19. 如权利要求15所述的可读存储介质,其中,所述通过所述预设识别模型中的各辅助预测模块,根据所述编码向量对所述第一面试样本数据进行辅助标签预测,得到与各所述辅助预测模块输出的辅助标签分布,包括:The readable storage medium according to claim 15, wherein the auxiliary label prediction is performed on the first interview sample data according to the coding vector through each auxiliary prediction module in the preset recognition model, and the result is obtained with The auxiliary label distribution output by each of the auxiliary prediction modules, including:
    获取所述面试编码向量中的至少两个第二正向编码向量以及至少两个第二反向编码向量;所述第二正向编码是指对各未标注样本字词按照正向顺序进行编码得到;所述第二反向编码是指对各所述未标注样本字词按照反向顺序进行编码得到;Obtain at least two second forward encoding vectors and at least two second reverse encoding vectors in the interview encoding vectors; the second forward encoding refers to encoding each unlabeled sample word in a forward order Obtained; the second reverse encoding refers to encoding each of the unlabeled sample words in reverse order to obtain;
    根据各所述第二正向编码向量确定与各所述未标注样本字词对应的各正向辅助标签分布;同时,根据所述第二反向编码向量确定与各所述未标注样本字词对应的反向辅助标签分布。According to each of the second forward coding vectors, determine the distribution of each forward auxiliary label corresponding to each of the unlabeled sample words; at the same time, according to the second reverse coding vector The corresponding reverse auxiliary label distribution.
  20. 一个或多个存储有计算机可读指令的可读存储介质,其中,所述计算机可读指令被一个或多个处理器执行时,使得所述一个或多个处理器执行如下步骤:One or more readable storage media storing computer-readable instructions, wherein the computer-readable instructions, when executed by one or more processors, cause the one or more processors to perform the following steps:
    获取目标面试者的面试信息;所述面试信息中包含至少一个面试句子;一个所述面试句子中包含多个面试信息字词;Obtain the interview information of the target interviewee; the interview information includes at least one interview sentence; one of the interview sentences includes a plurality of interview information words;
    将所述面试句子输入至面试实体识别模型中,对所述面试句子中的面试信息字词进行提取识别,得到与各所述面试信息字词对应的实体识别结果;所述面试实体识别模型是根据如权利要求1至5任一项所述面试实体识别模型训练方法得到的;The interview sentence is input into the interview entity recognition model, and the interview information words in the interview sentence are extracted and recognized to obtain entity recognition results corresponding to each of the interview information words; the interview entity recognition model is Obtained according to the interview entity recognition model training method as described in any one of claims 1 to 5;
    将所述实体识别结果按照预设匹配规则插入预设面试信息存储模板。Inserting the entity recognition result into a preset interview information storage template according to preset matching rules.
PCT/CN2021/096583 2020-12-30 2021-05-28 Method and apparatus for training interview entity recognition model, and method and apparatus for extracting interview information entity WO2022142108A1 (en)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
CN202011620124.1A CN112733539A (en) 2020-12-30 2020-12-30 Interview entity recognition model training and interview information entity extraction method and device
CN202011620124.1 2020-12-30

Publications (1)

Publication Number Publication Date
WO2022142108A1 true WO2022142108A1 (en) 2022-07-07

Family

ID=75608490

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2021/096583 WO2022142108A1 (en) 2020-12-30 2021-05-28 Method and apparatus for training interview entity recognition model, and method and apparatus for extracting interview information entity

Country Status (2)

Country Link
CN (1) CN112733539A (en)
WO (1) WO2022142108A1 (en)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN115204320A (en) * 2022-09-15 2022-10-18 北京数牍科技有限公司 Naive Bayes model training method, device, equipment and computer storage medium

Families Citing this family (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN112733539A (en) * 2020-12-30 2021-04-30 平安科技(深圳)有限公司 Interview entity recognition model training and interview information entity extraction method and device
CN113434676B (en) * 2021-06-25 2023-12-22 平安国际智慧城市科技股份有限公司 Text relation extraction model training, text relation extraction method, device and equipment

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111062215A (en) * 2019-12-10 2020-04-24 金蝶软件(中国)有限公司 Named entity recognition method and device based on semi-supervised learning training
CN111860669A (en) * 2020-07-27 2020-10-30 平安科技(深圳)有限公司 Training method and device of OCR recognition model and computer equipment
CN112115267A (en) * 2020-09-28 2020-12-22 平安科技(深圳)有限公司 Training method, device and equipment of text classification model and storage medium
CN112733539A (en) * 2020-12-30 2021-04-30 平安科技(深圳)有限公司 Interview entity recognition model training and interview information entity extraction method and device

Family Cites Families (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US11003950B2 (en) * 2019-03-29 2021-05-11 Innoplexus Ag System and method to identify entity of data
CN111310823B (en) * 2020-02-12 2024-03-29 北京迈格威科技有限公司 Target classification method, device and electronic system
CN111737581A (en) * 2020-07-24 2020-10-02 网思分析(研究与技术)有限公司 Semi-supervised multi-task learning model for emotion analysis of specific aspect

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111062215A (en) * 2019-12-10 2020-04-24 金蝶软件(中国)有限公司 Named entity recognition method and device based on semi-supervised learning training
CN111860669A (en) * 2020-07-27 2020-10-30 平安科技(深圳)有限公司 Training method and device of OCR recognition model and computer equipment
CN112115267A (en) * 2020-09-28 2020-12-22 平安科技(深圳)有限公司 Training method, device and equipment of text classification model and storage medium
CN112733539A (en) * 2020-12-30 2021-04-30 平安科技(深圳)有限公司 Interview entity recognition model training and interview information entity extraction method and device

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN115204320A (en) * 2022-09-15 2022-10-18 北京数牍科技有限公司 Naive Bayes model training method, device, equipment and computer storage medium
CN115204320B (en) * 2022-09-15 2022-11-15 北京数牍科技有限公司 Naive Bayes model training method, device, equipment and computer storage medium

Also Published As

Publication number Publication date
CN112733539A (en) 2021-04-30

Similar Documents

Publication Publication Date Title
WO2022142108A1 (en) Method and apparatus for training interview entity recognition model, and method and apparatus for extracting interview information entity
WO2022142613A1 (en) Training corpus expansion method and apparatus, and intent recognition model training method and apparatus
CN111444723B (en) Information extraction method, computer device, and storage medium
CN107808011B (en) Information classification extraction method and device, computer equipment and storage medium
US11941366B2 (en) Context-based multi-turn dialogue method and storage medium
WO2020177230A1 (en) Medical data classification method and apparatus based on machine learning, and computer device and storage medium
WO2018153265A1 (en) Keyword extraction method, computer device, and storage medium
US11983493B2 (en) Data processing method and pronoun resolution neural network training method
WO2021212749A1 (en) Method and apparatus for labelling named entity, computer device, and storage medium
WO2022134805A1 (en) Document classification prediction method and apparatus, and computer device and storage medium
WO2022116436A1 (en) Text semantic matching method and apparatus for long and short sentences, computer device and storage medium
CN112380837B (en) Similar sentence matching method, device, equipment and medium based on translation model
WO2022141864A1 (en) Conversation intent recognition model training method, apparatus, computer device, and medium
CN110427612B (en) Entity disambiguation method, device, equipment and storage medium based on multiple languages
WO2021051574A1 (en) English text sequence labelling method and system, and computer device
US11170167B2 (en) Automatic lexical sememe prediction system using lexical dictionaries
CN112926654A (en) Pre-labeling model training and certificate pre-labeling method, device, equipment and medium
WO2022227162A1 (en) Question and answer data processing method and apparatus, and computer device and storage medium
WO2021189960A1 (en) Method and apparatus for training adversarial network, method and apparatus for supplementing medical data, and device and medium
JP7480811B2 (en) Method of sample analysis, electronic device, computer readable storage medium, and computer program product
CN112652295A (en) Language model training method, device, equipment and medium, and video subtitle checking method, device and medium
CN110633475A (en) Natural language understanding method, device and system based on computer scene and storage medium
WO2021068524A1 (en) Image matching method and apparatus, computer device, and storage medium
WO2023065635A1 (en) Named entity recognition method and apparatus, storage medium and terminal device
CN112580329B (en) Text noise data identification method, device, computer equipment and storage medium

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 21912869

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 21912869

Country of ref document: EP

Kind code of ref document: A1