CN116737894A - Intelligent robot service system based on model training - Google Patents
Intelligent robot service system based on model training Download PDFInfo
- Publication number
- CN116737894A CN116737894A CN202310646279.XA CN202310646279A CN116737894A CN 116737894 A CN116737894 A CN 116737894A CN 202310646279 A CN202310646279 A CN 202310646279A CN 116737894 A CN116737894 A CN 116737894A
- Authority
- CN
- China
- Prior art keywords
- feedback information
- feedback
- encoder
- decoder
- module
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Granted
Links
- 238000012549 training Methods 0.000 title claims abstract description 23
- 238000012512 characterization method Methods 0.000 claims abstract description 7
- 238000000605 extraction Methods 0.000 claims abstract description 6
- 239000013598 vector Substances 0.000 claims description 52
- 230000007246 mechanism Effects 0.000 claims description 40
- 238000000034 method Methods 0.000 claims description 27
- 230000008569 process Effects 0.000 claims description 27
- 230000006870 function Effects 0.000 claims description 25
- 238000013528 artificial neural network Methods 0.000 claims description 12
- 238000004364 calculation method Methods 0.000 claims description 11
- 238000005457 optimization Methods 0.000 claims description 9
- 230000002708 enhancing effect Effects 0.000 claims description 3
- 238000013507 mapping Methods 0.000 claims description 3
- 238000013473 artificial intelligence Methods 0.000 abstract description 5
- 230000004048 modification Effects 0.000 description 5
- 238000012986 modification Methods 0.000 description 5
- 230000000694 effects Effects 0.000 description 3
- 230000003993 interaction Effects 0.000 description 3
- 238000010586 diagram Methods 0.000 description 2
- 238000012804 iterative process Methods 0.000 description 2
- 238000012545 processing Methods 0.000 description 2
- 230000004044 response Effects 0.000 description 2
- 238000006467 substitution reaction Methods 0.000 description 2
- 230000009286 beneficial effect Effects 0.000 description 1
- 238000004891 communication Methods 0.000 description 1
- 238000011161 development Methods 0.000 description 1
- 238000005516 engineering process Methods 0.000 description 1
- 239000000284 extract Substances 0.000 description 1
- 230000006872 improvement Effects 0.000 description 1
- 238000002372 labelling Methods 0.000 description 1
- 230000007774 longterm Effects 0.000 description 1
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F16/00—Information retrieval; Database structures therefor; File system structures therefor
- G06F16/30—Information retrieval; Database structures therefor; File system structures therefor of unstructured textual data
- G06F16/33—Querying
- G06F16/332—Query formulation
- G06F16/3329—Natural language query formulation or dialogue systems
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F16/00—Information retrieval; Database structures therefor; File system structures therefor
- G06F16/30—Information retrieval; Database structures therefor; File system structures therefor of unstructured textual data
- G06F16/33—Querying
- G06F16/335—Filtering based on additional data, e.g. user or group profiles
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
- G06N3/045—Combinations of networks
- G06N3/0455—Auto-encoder networks; Encoder-decoder networks
Landscapes
- Engineering & Computer Science (AREA)
- Physics & Mathematics (AREA)
- Theoretical Computer Science (AREA)
- Computational Linguistics (AREA)
- Mathematical Physics (AREA)
- Data Mining & Analysis (AREA)
- General Engineering & Computer Science (AREA)
- General Physics & Mathematics (AREA)
- Artificial Intelligence (AREA)
- Databases & Information Systems (AREA)
- Life Sciences & Earth Sciences (AREA)
- Health & Medical Sciences (AREA)
- Human Computer Interaction (AREA)
- Biomedical Technology (AREA)
- Biophysics (AREA)
- Evolutionary Computation (AREA)
- General Health & Medical Sciences (AREA)
- Molecular Biology (AREA)
- Computing Systems (AREA)
- Software Systems (AREA)
- Electrically Operated Instructional Devices (AREA)
- Machine Translation (AREA)
Abstract
The invention relates to the field of artificial intelligence, in particular to an intelligent robot service system based on model training. The system comprises a selection module, a feedback module and a feedback module, wherein the selection module is used for selecting any problem to be fed back from a first database; the input module is used for inputting the problem to be fed back into a language model of a transducer architecture to form a first feedback list; the receiving module is used for receiving the feedback information of the user when the feedback information presents different information characterization sequences, so as to form a second feedback list; the extraction module is used for extracting first feedback information positioned at the first position in the first feedback list, comparing the first feedback information with the service satisfaction degree of the first feedback information in the second feedback list and obtaining a comparison result; and the adjustment module adjusts the adjustment strategy of the language model according to the comparison result so as to enable the first feedback information in the first feedback list to be the optimal feedback information. According to the invention, the language model is adjusted, so that the satisfaction degree of a person can be maximized in the language of the information fed back by the robot system for the problem raised by the user.
Description
Technical Field
The invention relates to the field of artificial intelligence, in particular to an intelligent robot service system based on model training.
Background
With the rapid development of science and technology, artificial intelligence has deepened into various industries, intelligent machines in some industries can search and answer simple questions presented by users, and the machines can answer the questions but still have limitations, especially in some cases, the answers which users want to get are more temperature, so that in order to optimize the answers of the artificial intelligence questions, the interactivity with people is improved, and an artificial intelligence optimizing system needs to be designed.
Disclosed in patent document publication No. CN115759123a is an intelligent question-answering robot system, which includes an acquisition model for acquiring a question posed by a user, a question-answering model, and an output model; the question-answering model is used for finding out the question which is most matched with the user query, and further giving out a corresponding answer, and the question-answering model adopts two forms of converting words into vectors and word shift distances; the output model is used for receiving the corresponding answers given by the question-answer model and outputting the answers.
The intelligent question-answering robot in the prior art answers questions presented by users by searching a database, and feedback information presented by the intelligent question-answering robot is single in form and has limitations and poor interaction effect.
Disclosure of Invention
Therefore, the intelligent robot service system based on model training can solve the problems that the feedback information form of a robot is single, limitation exists, and the interaction effect is poor.
To achieve the above object, the present invention provides an intelligent robot service system based on model training, comprising: the selection module is used for selecting any problem to be fed back from a first database, and a plurality of problems to be fed back are prestored in the first database;
the input module is connected with the selection module and used for inputting the to-be-fed-back problem into a language model based on a transducer architecture, outputting at least one piece of feedback information based on the to-be-fed-back problem, wherein a plurality of pieces of feedback information form a first feedback list, and the feedback information comprises at least two information characterizations;
the receiving module is used for receiving user feedback information when the feedback information presents different information characterization sequences, wherein the user feedback information is used for representing service satisfaction degree of a user on the feedback information, and ordering the feedback information from high to low based on the service satisfaction degree to form a second feedback list;
the extraction module is connected with the input module and used for extracting first feedback information positioned at the first position in the first feedback list, comparing the first feedback information with the service satisfaction degree of the first feedback information in the second feedback list and obtaining a comparison result;
the adjusting module is connected with the extracting module and used for adjusting the adjusting strategy of the language model of the transducer architecture according to the comparison result so as to enable the first feedback information in the first feedback list to be the optimal feedback information.
Further, the adjustment module comprises a joining unit, a calculating unit and an iteration unit;
the adding unit is used for adding KL divergence into a loss function of a language model of the transducer architecture as one item of the loss function;
the calculation unit is used for calculating the KL divergence of the loss function according to the comparison result, wherein the KL divergence r KL The expression of (2) isSaid pi RL Output distribution probability representing the first feedback information of the language model outputting the transducer architecture, the pi SFT The output distribution probability of the feedback information after the adjustment module adjusts the language model of the transducer architecture is represented;
the iteration unit is connected with the calculation unit and used for enabling the first feedback information in the first feedback list to be iterated into optimal feedback information by minimizing the KL divergence.
Further, in the iterative process, the iteration unit is preset with a standard threshold, and when a comparison result is obtained, if the service satisfaction degree of the first feedback information in the first feedback list and the first feedback information in the second feedback list is greater than or equal to the standard threshold, the KL divergence value is reduced for multiple times, so that the language model of the transducer architecture is updated for multiple times, and the first feedback information in the first feedback list is output to the optimal feedback information until the KL divergence value is minimum.
Further, if the service satisfaction degree of the first feedback information in the first feedback list and the first feedback information in the second feedback list is smaller than the standard threshold value, updating the language model of the Transformer architecture once to enable the KL divergence value to be minimum, and outputting optimal feedback information by the first feedback information in the first feedback list.
Further, the language model of the transducer architecture includes: an input embedded layer, an encoder, a decoder, and an output layer;
the input embedding layer is used for converting word sequences input into questions to be fed back into vector representations;
the encoder is configured to convert the vector representation into an output vector of the encoder;
the decoder is used for converting the output vector of the encoder into a query vector, and calculating the query vector and the output vector representation of the encoder to obtain the context representation of the problem to be fed back to obtain the output vector of the decoder;
the output layer is used for mapping the output vector of the decoder into the answer of the feedback question.
Further, the encoder comprises an encoder position encoder, an encoder multi-head self-attention mechanism and an encoder forward neural network layer, and the working process of the encoder is as follows:
encoding position information of words in each of the questions to be fed back into a vector by using the encoder position encoder;
inputting the vector into the multi-head self-attention mechanism of the encoder, and establishing the relation between words in the to-be-fed-back problem in an input sequence;
and using the encoder forward neural network to process the representation of the word in each to-be-fed-back problem in the multi-head self-attention mechanism of the encoder to obtain an output vector of the encoder.
Further, the decoder comprises a decoder position encoder, a decoder multi-head self-attention mechanism, a multi-head attention mechanism and a decoder forward neural network layer, and the working process of the decoder is as follows:
encoding the position information of the words in each of the questions to be fed back into a vector by using the decoder position encoder;
inputting the vector into a multi-head self-attention mechanism of a decoder, and establishing a relation between words in the decoder to obtain an intermediate layer representation of the decoder;
inputting an intermediate layer representation of the decoder into a multi-headed attention mechanism;
the decoder forward neural network is used to process the representation of each word in the multi-headed attention mechanism to obtain the output vector of the decoder.
The working process of the language model of the transducer architecture is as follows:
inputting word sequences in a problem to be fed back into the input embedding layer to be converted into vector representations;
the multi-head self-attention mechanism of the encoder processes the words of the problem to be fed back, and calculates the importance degree of each word of the problem to be fed back to the sentence of the problem to be fed back according to the current position and semantic relation of the word of the problem to be fed back, so as to obtain the representation of the word of the problem to be fed back in the input process;
encoding the representation of the to-be-fed-back problem word obtained by the multi-head self-attention mechanism of the encoder in input through the feedforward full-connection layer, and enhancing the semantic expression capability of the representation;
taking output vectors passing through the encoder multi-head self-attention mechanism and the feedforward full-connection layer as input, and outputting answers to feedback questions through the encoder self-attention mechanism and the multi-layer decoder;
and obtaining the probability distribution of the answer of the output feedback problem by using a normalized exponential function.
Further, the receiving module counts the feedback information of the user through big data, gives the service satisfaction degree of the user on the feedback information based on the statistics result, and sorts the feedback information based on the service satisfaction degree from high to low.
Further, the minimized KL divergence is obtained by a gradient optimization algorithm, and the gradient optimization algorithm is used for searching the minimum value of the function by calculating the gradient of the objective function, by continuously adjusting the function parameters, so that the value of the objective function is continuously reduced until a local minimum value or a global minimum value is reached.
Compared with the prior art, the intelligent robot service system based on model training has the beneficial effects that the intelligent robot service system based on model training comprises the selection module, the input module, the receiving module, the extraction module and the adjustment module, wherein the selection module reduces the repeatability and improves the efficiency of the subsequent working process by selecting any problem to be fed back from the first database; the input module outputs at least one feedback message from any to-be-fed back problem selected from the first database through a language model based on a transform architecture, and the language model of the transform architecture has parallel computing capability, so that the consumption of computing resources is reduced; the receiving module sorts the feedback information from high to low based on the service satisfaction, so that scoring consistency of labeling personnel is improved; the extraction module is used for extracting first feedback information positioned at the first position in the first feedback list and comparing the service satisfaction degree of the first feedback information in the second feedback list to obtain a comparison result; the language model of the transducer architecture is adjusted through the adjusting module, so that the satisfaction degree of people can be maximized when the robot system feeds back information of the problems raised by the users.
Further, the adjustment module comprises a joining unit, a calculating unit and an iteration unit, and the adjustment module can continuously update the language model of the transducer architecture, so that the answer output by the language model of the transducer architecture is close to the answer with highest satisfaction.
Further, the iteration unit sets a standard threshold, and the updating process of the language model of the converter architecture can be judged through the standard threshold, so that the updating efficiency of the language model of the converter architecture is improved, and the optimal updating result of the language model of the converter architecture is realized.
Further, the language model of the transducer architecture has the advantages of no limitation on position-related operation, strong modeling capability, strong universality and strong expandability.
Further, the language model of the transducer architecture through the multi-head self-attention mechanism can focus on different parts of the problem to be fed back from different positions and angles, and the representation capability of the model is improved.
Further, the gradient optimization algorithm can select a reasonable parameter updating direction, and efficiency of KL divergence minimization is improved.
Drawings
FIG. 1 is a schematic diagram of an intelligent robot service system based on model training according to an embodiment of the present invention;
fig. 2 is a schematic diagram of another structure of an intelligent robot service system based on model training according to an embodiment of the present invention.
Detailed Description
In order that the objects and advantages of the invention will become more apparent, the invention will be further described with reference to the following examples; it should be understood that the specific embodiments described herein are for purposes of illustration only and are not intended to limit the scope of the invention.
Preferred embodiments of the present invention are described below with reference to the accompanying drawings. It should be understood by those skilled in the art that these embodiments are merely for explaining the technical principles of the present invention, and are not intended to limit the scope of the present invention.
It should be noted that, in the description of the present invention, terms such as "upper," "lower," "left," "right," "inner," "outer," and the like indicate directions or positional relationships based on the directions or positional relationships shown in the drawings, which are merely for convenience of description, and do not indicate or imply that the apparatus or elements must have a specific orientation, be constructed and operated in a specific orientation, and thus should not be construed as limiting the present invention.
Furthermore, it should be noted that, in the description of the present invention, unless explicitly specified and limited otherwise, the terms "mounted," "connected," and "connected" are to be construed broadly, and may be either fixedly connected, detachably connected, or integrally connected, for example; can be mechanically or electrically connected; can be directly connected or indirectly connected through an intermediate medium, and can be communication between two elements. The specific meaning of the above terms in the present invention can be understood by those skilled in the art according to the specific circumstances.
Referring to fig. 1, the present invention provides an intelligent robot service system based on model training, which includes:
the selection module 10 is configured to select any problem to be fed back from a first database, where a plurality of problems to be fed back are stored in the first database in advance;
the input module 20 is connected with the selection module 10, and is configured to input the to-be-fed-back problem to a language model based on a transducer architecture, output at least one feedback information based on the to-be-fed-back problem, and form a first feedback list by a plurality of feedback information, where the feedback information includes at least two information characterizations;
the receiving module 30 is configured to receive user feedback information when the feedback information presents different information characterization sequences, where the user feedback information is used to represent service satisfaction of a user on the feedback information, and order the feedback information from high to low based on the service satisfaction, so as to form a second feedback list;
the extracting module 40 is connected with the input module 20, and is configured to extract first feedback information located at the first position in the first feedback list, and compare the first feedback information with service satisfaction of the first feedback information in the second feedback list, so as to obtain a comparison result;
the adjustment module 50 is connected to the extraction module 40, and is configured to adjust an adjustment policy of the language model of the transducer architecture according to the comparison result, so that the first feedback information in the first feedback list is the optimal feedback information.
Specifically, the selection module 10 selects any problem to be fed back from the first database, so that the selected problem reduces repeatability and improves efficiency of the subsequent working process; the input module 20 outputs at least one feedback message from any problem to be fed back selected from the first database through a language model based on a transform architecture, and the language model of the transform architecture has parallel computing capability, so that the consumption of computing resources is reduced; the receiving module 30 orders the feedback information from high to low based on the service satisfaction, improving consistency of the satisfaction; the extracting module 40 extracts first feedback information in the first feedback list and service satisfaction of the first feedback information in the second feedback list to obtain a comparison result and help subsequent calculation; the language model of the transducer architecture is adjusted by the adjusting module 50, so that the satisfaction degree of people can be highest when the robot system feeds back information to the problem posed by the user, better optimization of feedback information output in the human-computer interaction process is realized, and the practicability of the feedback information is improved.
Specifically, referring to fig. 2, the adjustment module 50 includes a joining unit 51, a calculating unit 52, and an iterating unit 53;
the adding unit 51 is configured to add KL divergence as one of the loss functions in the loss function of the language model of the transducer architecture;
the calculation unit 52 is configured to calculate a KL divergence of the loss function according to the comparison result, wherein the KL divergence r KL The expression of (2) isThe saidπ RL Output distribution probability representing the first feedback information of the language model outputting the transducer architecture, the pi SFT The output distribution probability of the feedback information after the adjustment module 50 adjusts the language model of the transducer architecture is represented;
the iteration unit 53 is connected to the calculation unit 52 for iterating the first feedback information in the first feedback list to the optimal feedback information by minimizing the KL divergence.
Specifically, the adding unit 51 adds KL divergence to the loss function in the language model of the transducer architecture to help update the language model of the transducer architecture, so that the language model of the transducer architecture is updated with minimum modification, and the output calculation result of the first feedback information and the feedback information after the adjustment of the language model of the transducer architecture is close by calculating KL divergence.
Specifically, in the iterative process, the iterative unit 53 sets a standard threshold in advance, and when obtaining the comparison result, if the service satisfaction degree of the first feedback information in the first feedback list and the first feedback information in the second feedback list is greater than or equal to the standard threshold, the KL divergence value is reduced multiple times, so as to implement multiple updates on the language model of the transducer architecture, until the KL divergence value is minimum, the first feedback information in the first feedback list is output with optimal feedback information.
Specifically, the updating process of the language model of the transducer architecture can be judged through the standard threshold, the updating efficiency of the language model of the transducer architecture is improved, the optimal updating result of the language model of the transducer architecture is realized, further effective optimization of feedback information is realized, and the optimizing efficiency is improved.
Specifically, if the service satisfaction degree of the first feedback information in the first feedback list and the first feedback information in the second feedback list is smaller than the standard threshold value, updating the language model of the Transformer architecture once to enable the KL divergence value to be minimum, and outputting optimal feedback information by the first feedback information in the first feedback list.
Specifically, when the service satisfaction degree of the first feedback information in the first feedback list and the first feedback information in the second feedback list is smaller than the standard threshold, the language model updating process of the transducer architecture is reduced, and the language model updating efficiency of the transducer architecture is improved.
Specifically, the language model of the transducer architecture includes: an input embedded layer, an encoder, a decoder, and an output layer;
the input embedding layer is used for converting word sequences input into questions to be fed back into vector representations;
the encoder is configured to convert the vector representation into an output vector of the encoder;
the decoder is used for converting the output vector of the encoder into a query vector, and calculating the query vector and the output vector representation of the encoder to obtain the context representation of the problem to be fed back to obtain the output vector of the decoder;
the output layer is used for mapping the output vector of the decoder into the answer of the feedback question.
Specifically, the structure of the language model of the transducer architecture has the advantages of realizing complete parallel calculation, being flexible and extensible in structure, good in pre-training effect and processing multi-mode data.
Specifically, the encoder comprises an encoder position encoder, an encoder multi-head self-attention mechanism and an encoder forward neural network layer, and the working process of the encoder is as follows:
encoding position information of words in each of the questions to be fed back into a vector by using the encoder position encoder;
inputting the vector into the multi-head self-attention mechanism of the encoder, and establishing the relation between words in the to-be-fed-back problem in an input sequence;
and using the encoder forward neural network to process the representation of the word in each to-be-fed-back problem in the multi-head self-attention mechanism of the encoder to obtain an output vector of the encoder.
Specifically, the encoder adopts the encoder self-attention mechanism, so that the model can process information of all positions in the input sequence simultaneously in one calculation, parallelism and efficiency of the model are improved, and the encoder simultaneously uses the encoder multi-head self-attention mechanism, so that the capability of the model to learn relations among different positions in the input sequence is improved.
Specifically, the decoder comprises a decoder position encoder, a decoder multi-head self-attention mechanism, a multi-head attention mechanism and a decoder forward neural network layer, and the working process of the decoder is as follows:
encoding the position information of the words in each of the questions to be fed back into a vector by using the decoder position encoder;
inputting the vector into a multi-head self-attention mechanism of a decoder, and establishing a relation between words in the decoder to obtain an intermediate layer representation of the decoder;
inputting an intermediate layer representation of the decoder into a multi-headed attention mechanism;
the decoder forward neural network is used to process the representation of each word in the multi-headed attention mechanism to obtain the output vector of the decoder.
In particular, the decoder uses the decoder multi-headed self-attention mechanism to improve the modeling capability of the model so that it is more accurate in processing the input sequence.
Specifically, the working process of the language model of the transducer architecture is as follows:
inputting word sequences in a problem to be fed back into the input embedding layer to be converted into vector representations;
the multi-head self-attention mechanism of the encoder processes the words of the problem to be fed back, and calculates the importance degree of each word of the problem to be fed back to the sentence of the problem to be fed back according to the current position and semantic relation of the word of the problem to be fed back, so as to obtain the representation of the word of the problem to be fed back in the input process;
encoding the representation of the to-be-fed-back problem word obtained by the multi-head self-attention mechanism of the encoder in input through the feedforward full-connection layer, and enhancing the semantic expression capability of the representation;
taking output vectors passing through the encoder multi-head self-attention mechanism and the feedforward full-connection layer as input, and outputting answers to feedback questions through the encoder self-attention mechanism and the multi-layer decoder;
and obtaining the probability distribution of the answer of the output feedback problem by using a normalized exponential function.
Specifically, the language model of the transducer architecture has strong context awareness capability when analyzing text and sequence data, so that the model has long-term dependence in learning text and sequence, the structure of the language model of the transducer architecture has high efficiency, parallelism and expandability, and the reasoning speed of the model is fast for a large-scale data set.
Specifically, the receiving module 30 counts the feedback information of the user through big data, gives the service satisfaction degree of the user to the feedback information based on the statistics result, and orders the feedback information from high to low based on the service satisfaction degree.
Specifically, the user feedback information data requires a large amount of data, so that the user satisfaction degree ranking reflected based on the statistical result of the user feedback information is more accurate.
Specifically, the minimized KL divergence is obtained by a gradient optimization algorithm, and the gradient optimization algorithm is used for searching the minimum value of the function by calculating the gradient of the objective function, by continuously adjusting the function parameters, so that the value of the objective function is continuously reduced until a local minimum value or a global minimum value is reached.
Specifically, the gradient optimization algorithm can select a reasonable parameter updating direction, and the KL divergence minimization efficiency is improved.
Specifically, the intelligent robot service system based on model training provided by the embodiment of the invention can be used for natural language understanding and generation of the intelligent question-answering robot for credit service, realizes accurate understanding and quick response to various problems of users in the field of credit service, reduces the workload of manual customer service, improves the response speed and accuracy of customer consultation, and provides more professional and convenient financial service for the users.
Thus far, the technical solution of the present invention has been described in connection with the preferred embodiments shown in the drawings, but it is easily understood by those skilled in the art that the scope of protection of the present invention is not limited to these specific embodiments. Equivalent modifications and substitutions for related technical features may be made by those skilled in the art without departing from the principles of the present invention, and such modifications and substitutions will be within the scope of the present invention.
The foregoing description is only of the preferred embodiments of the invention and is not intended to limit the invention; various modifications and variations of the present invention will be apparent to those skilled in the art. Any modification, equivalent replacement, improvement, etc. made within the spirit and principle of the present invention should be included in the protection scope of the present invention.
Claims (10)
1. An intelligent robot service system based on model training, comprising:
the selection module is used for selecting any problem to be fed back from a first database, and a plurality of problems to be fed back are prestored in the first database;
the input module is connected with the selection module and used for inputting the to-be-fed-back problem into a language model based on a transducer architecture, outputting at least one piece of feedback information based on the to-be-fed-back problem, wherein a plurality of pieces of feedback information form a first feedback list, and the feedback information comprises at least two information characterizations;
the receiving module is used for receiving user feedback information when the feedback information presents different information characterization sequences, wherein the user feedback information is used for representing service satisfaction degree of a user on the feedback information, and ordering the feedback information from high to low based on the service satisfaction degree to form a second feedback list;
the extraction module is connected with the input module and used for extracting first feedback information positioned at the first position in the first feedback list, comparing the first feedback information with the service satisfaction degree of the first feedback information in the second feedback list and obtaining a comparison result;
the adjusting module is connected with the extracting module and used for adjusting the adjusting strategy of the language model of the transducer architecture according to the comparison result so as to enable the first feedback information in the first feedback list to be the optimal feedback information.
2. The model training-based intelligent robot service system of claim 1, wherein the adjustment module comprises a joining unit, a calculation unit, and an iteration unit;
the adding unit is used for adding KL divergence into a loss function of a language model of the transducer architecture as one item of the loss function;
the calculation unit is used for calculating the KL divergence of the loss function according to the comparison result, wherein the KL divergence r KL The expression of (2) isSaid pi RL Output distribution probability representing the first feedback information of the language model outputting the transducer architecture, the pi SFT The output distribution probability of the feedback information after the adjustment module adjusts the language model of the transducer architecture is represented;
the iteration unit is connected with the calculation unit and used for enabling the first feedback information in the first feedback list to be iterated into optimal feedback information by minimizing the KL divergence.
3. The intelligent robot service system based on model training according to claim 2, wherein the iteration unit presets a standard threshold in the process of iteration, and when a comparison result is obtained, if the service satisfaction degree of the first feedback information in the first feedback list and the first feedback information in the second feedback list is greater than or equal to the standard threshold, the KL divergence value is reduced for multiple times, so as to update the language model of the transducer architecture for multiple times, and the first feedback information in the first feedback list is output to the optimal feedback information until the KL divergence value is minimum.
4. The intelligent robot service system based on model training according to claim 3, wherein if the service satisfaction degree of the first feedback information in the first feedback list and the first feedback information in the second feedback list is smaller than the standard threshold value, the language model of the transducer architecture is updated once so that the KL divergence value becomes minimum, and the first feedback information in the first feedback list outputs the optimal feedback information.
5. The model training-based intelligent robot service system of claim 4, wherein the language model of the transducer architecture comprises: an input embedded layer, an encoder, a decoder, and an output layer;
the input embedding layer is used for converting word sequences input into questions to be fed back into vector representations;
the encoder is configured to convert the vector representation into an output vector of the encoder;
the decoder is used for converting the output vector of the encoder into a query vector, and calculating the query vector and the output vector representation of the encoder to obtain the context representation of the problem to be fed back to obtain the output vector of the decoder;
the output layer is used for mapping the output vector of the decoder into the answer of the feedback question.
6. The model training-based intelligent robot service system of claim 5, wherein the encoder comprises an encoder position encoder, an encoder multi-headed self-attention mechanism and an encoder forward neural network layer, the encoder working process is as follows:
encoding position information of words in each of the questions to be fed back into a vector by using the encoder position encoder;
inputting the vector into the multi-head self-attention mechanism of the encoder, and establishing the relation between words in the to-be-fed-back problem in an input sequence;
and using the encoder forward neural network to process the representation of the word in each to-be-fed-back problem in the multi-head self-attention mechanism of the encoder to obtain an output vector of the encoder.
7. The model training based intelligent robotic service system of claim 6, wherein the decoder comprises a decoder position encoder, a decoder multi-headed self-attention mechanism, a multi-headed attention mechanism, and a decoder forward neural network layer, the decoder operating as:
encoding the position information of the words in each of the questions to be fed back into a vector by using the decoder position encoder;
inputting the vector into a multi-head self-attention mechanism of a decoder, and establishing a relation between words in the decoder to obtain an intermediate layer representation of the decoder;
inputting an intermediate layer representation of the decoder into a multi-headed attention mechanism;
the decoder forward neural network is used to process the representation of each word in the multi-headed attention mechanism to obtain the output vector of the decoder.
8. The intelligent robot service system based on model training of claim 7, wherein the language model of the transducer architecture works as follows:
inputting word sequences in a problem to be fed back into the input embedding layer to be converted into vector representations;
the multi-head self-attention mechanism of the encoder processes the words of the problem to be fed back, and calculates the importance degree of each word of the problem to be fed back to the sentence of the problem to be fed back according to the current position and semantic relation of the word of the problem to be fed back, so as to obtain the representation of the word of the problem to be fed back in the input process;
encoding the representation of the to-be-fed-back problem word obtained by the multi-head self-attention mechanism of the encoder in input through the feedforward full-connection layer, and enhancing the semantic expression capability of the representation;
taking output vectors passing through the encoder multi-head self-attention mechanism and the feedforward full-connection layer as input, and outputting answers to feedback questions through the encoder self-attention mechanism and the multi-layer decoder;
and obtaining the probability distribution of the answer of the output feedback problem by using a normalized exponential function.
9. The intelligent robot service system based on model training according to claim 8, wherein the receiving module counts the feedback information of the user through big data, gives the service satisfaction of the user for the feedback information based on the statistics result, and orders the feedback information from high to low based on the service satisfaction.
10. The intelligent robot service system based on model training according to claim 9, wherein the minimized KL divergence is obtained by a gradient optimization algorithm that finds the minimum of the function by calculating the gradient of the objective function by continuously adjusting the function parameters such that the value of the objective function is continuously reduced until a local minimum or global minimum is reached.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202310646279.XA CN116737894B (en) | 2023-06-02 | 2023-06-02 | Intelligent robot service system based on model training |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202310646279.XA CN116737894B (en) | 2023-06-02 | 2023-06-02 | Intelligent robot service system based on model training |
Publications (2)
Publication Number | Publication Date |
---|---|
CN116737894A true CN116737894A (en) | 2023-09-12 |
CN116737894B CN116737894B (en) | 2024-02-20 |
Family
ID=87912554
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN202310646279.XA Active CN116737894B (en) | 2023-06-02 | 2023-06-02 | Intelligent robot service system based on model training |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN116737894B (en) |
Citations (9)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20170124497A1 (en) * | 2015-10-28 | 2017-05-04 | Fractal Industries, Inc. | System for automated capture and analysis of business information for reliable business venture outcome prediction |
US20190251168A1 (en) * | 2018-02-09 | 2019-08-15 | Salesforce.Com, Inc. | Multitask Learning As Question Answering |
CN111444311A (en) * | 2020-02-26 | 2020-07-24 | 平安科技(深圳)有限公司 | Semantic understanding model training method and device, computer equipment and storage medium |
CN111881279A (en) * | 2020-07-28 | 2020-11-03 | 平安科技(深圳)有限公司 | Transformer model-based question answering method, question answering device and storage device |
CN112328767A (en) * | 2020-11-11 | 2021-02-05 | 重庆邮电大学 | Question-answer matching method based on BERT model and comparative aggregation framework |
US20210073215A1 (en) * | 2019-09-05 | 2021-03-11 | Verizon Patent And Licensing Inc. | Natural language-based content system with corrective feedback and training |
CN113032545A (en) * | 2021-05-29 | 2021-06-25 | 成都晓多科技有限公司 | Method and system for conversation understanding and answer configuration based on unsupervised conversation pre-training |
US20220108688A1 (en) * | 2020-10-02 | 2022-04-07 | Salesforce.Com, Inc. | Systems and methods for a multilingual speech recognition framework |
WO2023273170A1 (en) * | 2021-06-30 | 2023-01-05 | 同济人工智能研究院(苏州)有限公司 | Welcoming robot conversation method |
-
2023
- 2023-06-02 CN CN202310646279.XA patent/CN116737894B/en active Active
Patent Citations (9)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20170124497A1 (en) * | 2015-10-28 | 2017-05-04 | Fractal Industries, Inc. | System for automated capture and analysis of business information for reliable business venture outcome prediction |
US20190251168A1 (en) * | 2018-02-09 | 2019-08-15 | Salesforce.Com, Inc. | Multitask Learning As Question Answering |
US20210073215A1 (en) * | 2019-09-05 | 2021-03-11 | Verizon Patent And Licensing Inc. | Natural language-based content system with corrective feedback and training |
CN111444311A (en) * | 2020-02-26 | 2020-07-24 | 平安科技(深圳)有限公司 | Semantic understanding model training method and device, computer equipment and storage medium |
CN111881279A (en) * | 2020-07-28 | 2020-11-03 | 平安科技(深圳)有限公司 | Transformer model-based question answering method, question answering device and storage device |
US20220108688A1 (en) * | 2020-10-02 | 2022-04-07 | Salesforce.Com, Inc. | Systems and methods for a multilingual speech recognition framework |
CN112328767A (en) * | 2020-11-11 | 2021-02-05 | 重庆邮电大学 | Question-answer matching method based on BERT model and comparative aggregation framework |
CN113032545A (en) * | 2021-05-29 | 2021-06-25 | 成都晓多科技有限公司 | Method and system for conversation understanding and answer configuration based on unsupervised conversation pre-training |
WO2023273170A1 (en) * | 2021-06-30 | 2023-01-05 | 同济人工智能研究院(苏州)有限公司 | Welcoming robot conversation method |
Also Published As
Publication number | Publication date |
---|---|
CN116737894B (en) | 2024-02-20 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN112015868B (en) | Question-answering method based on knowledge graph completion | |
CN117171333B (en) | Electric power file question-answering type intelligent retrieval method and system | |
CN110287482B (en) | Semi-automatic participle corpus labeling training device | |
WO2023272748A1 (en) | Academic accurate recommendation-oriented heterogeneous scientific research information integration method and system | |
CN116150335A (en) | Text semantic retrieval method under military scene | |
CN112115242A (en) | Intelligent customer service question-answering system based on naive Bayes classification algorithm | |
CN116910086B (en) | Database query method and system based on self-attention syntax sensing | |
CN116341562A (en) | Similar problem generation method based on Unilm language model | |
US20240037335A1 (en) | Methods, systems, and media for bi-modal generation of natural languages and neural architectures | |
CN118093829A (en) | Legal consultation method based on knowledge graph enhancement | |
CN117828105A (en) | Electric power defect text knowledge graph construction method based on joint extraction | |
CN117349319A (en) | Database query optimization method based on deep reinforcement learning | |
CN117453897B (en) | Document question-answering method and system based on large model and genetic algorithm | |
CN116737894B (en) | Intelligent robot service system based on model training | |
CN117609436A (en) | College scientific research management question-answering system combining knowledge graph and large language model | |
CN117290478A (en) | Knowledge graph question-answering method, device, equipment and storage medium | |
CN117033579A (en) | Novel knowledge base system and method based on LLMS | |
CN116127956A (en) | Self-adaptive term normalization method based on double-tower model | |
KR20200047272A (en) | Indexing system and method using variational recurrent autoencoding | |
CN115238705A (en) | Semantic analysis result reordering method and system | |
CN111753053B (en) | Reading understanding improvement method based on pre-training model | |
CN116401356B (en) | Knowledge graph multi-round question-answering method and system based on historical information tracking | |
CN113590921B (en) | Information retrieval method and system based on interactive social contact | |
CN118484729B (en) | Scientific and technological achievement evaluation method and system based on innovation value analysis | |
CN118410152B (en) | Information processing method, question-answering method and question-answering system |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
GR01 | Patent grant | ||
GR01 | Patent grant |