WO2021238599A1 - 对话模型的训练方法、装置、计算机设备及存储介质 - Google Patents
对话模型的训练方法、装置、计算机设备及存储介质 Download PDFInfo
- Publication number
- WO2021238599A1 WO2021238599A1 PCT/CN2021/091954 CN2021091954W WO2021238599A1 WO 2021238599 A1 WO2021238599 A1 WO 2021238599A1 CN 2021091954 W CN2021091954 W CN 2021091954W WO 2021238599 A1 WO2021238599 A1 WO 2021238599A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- dialogue
- dialog
- feature
- reply
- model
- Prior art date
Links
- 238000000034 method Methods 0.000 title claims abstract description 88
- 238000012549 training Methods 0.000 title claims abstract description 63
- 230000004044 response Effects 0.000 claims abstract description 32
- 238000009826 distribution Methods 0.000 claims description 105
- 238000005070 sampling Methods 0.000 claims description 24
- 230000015654 memory Effects 0.000 claims description 16
- 238000012545 processing Methods 0.000 claims description 15
- 238000000605 extraction Methods 0.000 claims description 7
- 239000000284 extract Substances 0.000 claims description 6
- 238000013473 artificial intelligence Methods 0.000 abstract description 24
- 238000010586 diagram Methods 0.000 description 12
- 230000006870 function Effects 0.000 description 12
- 230000008569 process Effects 0.000 description 11
- 238000005516 engineering process Methods 0.000 description 10
- 230000002093 peripheral effect Effects 0.000 description 10
- 230000001133 acceleration Effects 0.000 description 9
- 238000004891 communication Methods 0.000 description 9
- 238000002474 experimental method Methods 0.000 description 9
- 230000003993 interaction Effects 0.000 description 8
- 238000010801 machine learning Methods 0.000 description 5
- 238000003058 natural language processing Methods 0.000 description 5
- 230000003287 optical effect Effects 0.000 description 5
- 238000013528 artificial neural network Methods 0.000 description 4
- 238000009795 derivation Methods 0.000 description 4
- 238000013461 design Methods 0.000 description 4
- 238000011156 evaluation Methods 0.000 description 4
- 238000013507 mapping Methods 0.000 description 4
- 230000000306 recurrent effect Effects 0.000 description 4
- 230000001965 increasing effect Effects 0.000 description 3
- 239000000919 ceramic Substances 0.000 description 2
- 230000000694 effects Effects 0.000 description 2
- 230000004927 fusion Effects 0.000 description 2
- 230000007246 mechanism Effects 0.000 description 2
- 238000005457 optimization Methods 0.000 description 2
- 238000013519 translation Methods 0.000 description 2
- NAWXUBYGYWOOIX-SFHVURJKSA-N (2s)-2-[[4-[2-(2,4-diaminoquinazolin-6-yl)ethyl]benzoyl]amino]-4-methylidenepentanedioic acid Chemical compound C1=CC2=NC(N)=NC(N)=C2C=C1CCC1=CC=C(C(=O)N[C@@H](CC(=C)C(O)=O)C(O)=O)C=C1 NAWXUBYGYWOOIX-SFHVURJKSA-N 0.000 description 1
- 241000282412 Homo Species 0.000 description 1
- 230000009471 action Effects 0.000 description 1
- 230000006399 behavior Effects 0.000 description 1
- 230000008901 benefit Effects 0.000 description 1
- 230000005540 biological transmission Effects 0.000 description 1
- 230000006835 compression Effects 0.000 description 1
- 238000007906 compression Methods 0.000 description 1
- 238000010276 construction Methods 0.000 description 1
- 230000007423 decrease Effects 0.000 description 1
- 230000003247 decreasing effect Effects 0.000 description 1
- 238000013135 deep learning Methods 0.000 description 1
- 238000011161 development Methods 0.000 description 1
- 230000009977 dual effect Effects 0.000 description 1
- 230000005484 gravity Effects 0.000 description 1
- 230000006872 improvement Effects 0.000 description 1
- 230000001939 inductive effect Effects 0.000 description 1
- 230000001788 irregular Effects 0.000 description 1
- 238000012804 iterative process Methods 0.000 description 1
- 239000004973 liquid crystal related substance Substances 0.000 description 1
- 239000000463 material Substances 0.000 description 1
- 238000005259 measurement Methods 0.000 description 1
- 230000005055 memory storage Effects 0.000 description 1
- 239000000203 mixture Substances 0.000 description 1
- 238000010295 mobile communication Methods 0.000 description 1
- 238000012986 modification Methods 0.000 description 1
- 230000004048 modification Effects 0.000 description 1
- 230000008447 perception Effects 0.000 description 1
- 230000009467 reduction Effects 0.000 description 1
- 230000002787 reinforcement Effects 0.000 description 1
- 238000009877 rendering Methods 0.000 description 1
- 238000011160 research Methods 0.000 description 1
- 230000006641 stabilisation Effects 0.000 description 1
- 238000011105 stabilization Methods 0.000 description 1
- 239000010409 thin film Substances 0.000 description 1
- 238000013526 transfer learning Methods 0.000 description 1
- 238000012795 verification Methods 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F16/00—Information retrieval; Database structures therefor; File system structures therefor
- G06F16/30—Information retrieval; Database structures therefor; File system structures therefor of unstructured textual data
- G06F16/33—Querying
- G06F16/332—Query formulation
- G06F16/3329—Natural language query formulation or dialogue systems
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F16/00—Information retrieval; Database structures therefor; File system structures therefor
- G06F16/30—Information retrieval; Database structures therefor; File system structures therefor of unstructured textual data
- G06F16/33—Querying
- G06F16/3331—Query processing
- G06F16/334—Query execution
- G06F16/3344—Query execution using natural language analysis
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F40/00—Handling natural language data
- G06F40/10—Text processing
- G06F40/12—Use of codes for handling textual entities
- G06F40/126—Character encoding
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F40/00—Handling natural language data
- G06F40/30—Semantic analysis
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
- G06N3/044—Recurrent networks, e.g. Hopfield networks
- G06N3/0442—Recurrent networks, e.g. Hopfield networks characterised by memory or gating, e.g. long short-term memory [LSTM] or gated recurrent units [GRU]
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
- G06N3/045—Combinations of networks
- G06N3/0455—Auto-encoder networks; Encoder-decoder networks
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
- G06N3/047—Probabilistic or stochastic networks
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/08—Learning methods
- G06N3/088—Non-supervised learning, e.g. competitive learning
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/08—Learning methods
- G06N3/098—Distributed learning, e.g. federated learning
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N7/00—Computing arrangements based on specific mathematical models
- G06N7/01—Probabilistic graphical models, e.g. probabilistic networks
-
- Y—GENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
- Y02—TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
- Y02D—CLIMATE CHANGE MITIGATION TECHNOLOGIES IN INFORMATION AND COMMUNICATION TECHNOLOGIES [ICT], I.E. INFORMATION AND COMMUNICATION TECHNOLOGIES AIMING AT THE REDUCTION OF THEIR OWN ENERGY USE
- Y02D10/00—Energy efficient computing, e.g. low power processors, power management or thermal management
Definitions
- This application relates to the field of artificial intelligence technology, and in particular to a method for training a dialogue model, a method for generating a dialogue reply, a device, a computer device, and a storage medium.
- natural language processing can be applied in a wider range.
- human-computer interaction scenarios such as small chat robots, dialogue systems, and terminal intelligent assistants.
- the computer device can output the corresponding dialogue reply according to the dialogue text input by the user during the dialogue process. How to avoid the overly monotonous dialog response output by computer equipment is a problem that needs to be solved.
- the embodiments of the application provide a method for training a dialogue model, a method for generating a dialogue reply, a device, a computer device, and a storage medium.
- a method for training a dialogue model By updating the parameters of the dialogue model multiple times according to the dialogue characteristics of the dialogue, the different semantics of the dialogue are taken into consideration, so that Dialogue replies contain multiple semantics, which improves the diversity of dialogue replies generated by the dialogue model.
- the technical solution is as follows:
- a method for training a dialogue model includes:
- the prior network is used to output the probability distribution of the dialog features
- the The posterior network is used to estimate the probability distribution of the dialogue features output by the prior network
- the first dialogue feature is used to represent the posterior features of the dialogue above and a dialogue reply in a dialogue
- the second dialogue feature A priori feature used to represent the dialogue above and a dialogue reply in a dialogue, the first dialogue including a first dialogue above and at least two first dialogue replies;
- the trained model is used as a dialogue model.
- a method for generating a dialogue reply includes:
- a training device for a dialogue model includes:
- a feature acquisition module for acquiring at least two first dialog features and at least two second dialog features of a first dialog based on a priori network and a posterior network in the dialog model, the prior network is used to output dialog features
- the posterior network is used to estimate the probability distribution of dialogue features output by the a priori network
- the first dialogue feature is used to represent the posterior features of a dialogue in a dialogue and a dialogue reply
- the second dialogue feature is used to represent a priori feature of the dialogue above and one dialogue reply in one dialogue
- the first dialogue includes a first dialogue above and at least two first dialogue replies;
- a model update module configured to update the dialogue model based on at least two first dialogue features and at least two second dialogue features of the first dialogue
- the model update module is further configured to update the posterior network based on at least two first dialog features of the first dialog;
- the model update module is further configured to update the discriminator of the dialogue model according to at least two first dialogue features and at least two second dialogue features of the second dialogue;
- the model acquisition module is used to respond to meeting the training end condition and use the trained model as a dialogue model.
- a device for generating a dialogue reply comprising:
- the dialogue acquisition module is used to acquire the dialogue above;
- a feature extraction module configured to input the dialogue above into a dialogue model, and randomly extract a target dialogue feature from the first dialogue features corresponding to a plurality of dialogue replies based on the prior network in the dialogue model;
- a reply output module configured to decode the target dialogue feature based on the decoder in the dialogue model, and output a target dialogue reply
- the reply display module is used to display the target dialogue reply.
- a computer device in another aspect, includes a processor and a memory, and the memory is used to store at least one piece of program code.
- the at least one piece of program code is loaded and executed by the processor to implement the present application. The operations performed in the training method of the dialog model in the embodiment, or the operations performed to realize the operations performed in the method for generating a dialog reply in the embodiment of the present application.
- a storage medium is provided, and at least one piece of program code is stored in the storage medium, and the at least one piece of program code is used to execute the dialog model training method in the embodiment of the present application, or to execute the embodiment of the present application.
- Fig. 1 is a schematic diagram of an implementation environment of a method for training a dialogue model according to an embodiment of the present application
- FIG. 2 is a flowchart of a method for training a dialogue model provided by an embodiment of the present application
- FIG. 3 is a flowchart of a method for generating a dialog reply provided by an embodiment of the present application
- FIG. 4 is a flowchart of a method for training a dialogue model provided by an embodiment of the present application
- Fig. 5 is a schematic structural diagram of a dialogue model provided according to an embodiment of the present application.
- Fig. 6 is a schematic flowchart of a multi-semantic WAE algorithm according to an embodiment of the present application
- Fig. 7 is a block diagram of a device for training a dialogue model according to an embodiment of the present application.
- Fig. 8 is a block diagram of a device for generating a dialog reply according to an embodiment of the present application.
- FIG. 9 is a structural block diagram of a terminal provided by an embodiment of the present application.
- Fig. 10 is a schematic structural diagram of a server provided according to an embodiment of the present application.
- AI Artificial Intelligence
- digital computers or machines controlled by digital computers to simulate, extend and expand human intelligence, perceive the environment, acquire knowledge, and use knowledge to obtain the best results.
- artificial intelligence is a comprehensive technology of computer science. It attempts to understand the essence of intelligence and produce a new kind of intelligent machine that can react in a similar way to human intelligence.
- Artificial intelligence is to study the design principles and implementation methods of various intelligent machines, so that the machines have the functions of perception, reasoning and decision-making.
- AIaaS Artificial intelligence cloud services are generally called AIaaS (AI as a Service, Chinese as "AI as a Service”).
- AIaaS Artificial intelligence platform service method
- the AIaaS platform will split several common AI services and provide independent or packaged services in the cloud.
- This service model is similar to opening an AI theme mall: all developers can access one or more artificial intelligence services provided by the platform through API interfaces, and some senior developers can also use the platform
- the AI framework and AI infrastructure provided are used to deploy and operate their own exclusive cloud artificial intelligence services.
- Natural language processing (Nature Language Processing, NLP) is an important direction in the field of computer science and artificial intelligence. It studies various theories and methods that can realize effective communication between humans and computers in natural language. Natural language processing is a science that integrates linguistics, computer science, and mathematics. Therefore, research in this field will involve natural language, that is, the language people use daily, so it is closely related to the study of linguistics. Natural language processing technologies usually include text processing, semantic understanding, machine translation, robot question answering, knowledge graphs and other technologies.
- Machine Learning is a multi-field interdisciplinary subject, involving probability theory, statistics, approximation theory, convex analysis, algorithm complexity theory and other subjects. Specializing in the study of how computers simulate or realize human learning behaviors in order to acquire new knowledge or skills, and reorganize the existing knowledge structure to continuously improve its own performance.
- Machine learning is the core of artificial intelligence, the fundamental way to make computers intelligent, and its applications cover all fields of artificial intelligence.
- Machine learning and deep learning usually include artificial neural networks, belief networks, reinforcement learning, transfer learning, inductive learning, teaching learning and other technologies.
- the embodiment of the present application provides a method for training a dialogue model, which can be implemented based on artificial intelligence technology.
- the dialogue model trained by this method can be applied to human-computer interaction scenarios. For example, chat robots, dialogue systems, and terminal intelligent assistants.
- chat robots When the user is chatting with the chat robot, the chat robot can input the content input by the user as the dialogue above into the dialogue model, and the dialogue model will output multiple dialogue replies, and then show the user one of the dialogue replies.
- the dialog system and the terminal intelligent assistant can also output a dialog reply that meets the user's needs based on the content input by the user.
- FIG. 1 is a schematic diagram of the implementation environment of the training method of the dialogue model provided according to an embodiment of the present application.
- the implementation environment may include: a terminal 110 and a server 120.
- the terminal 110 and the server 120 may be directly or indirectly connected through wired or wireless communication, which is not limited in this application.
- the terminal 110 may be a smart phone, a tablet computer, a notebook computer, a desktop computer, a smart speaker, a smart watch, etc., but is not limited thereto.
- the terminal 110 can install and run an application program that supports human-computer interaction.
- the application can be a chat robot application, a social application, a terminal intelligent assistant application, etc.
- the terminal 110 is a terminal used by a user, and a user account is logged in an application program running in the terminal 110.
- the server 120 may be an independent physical server, or a server cluster or distributed system composed of multiple physical servers, or it may provide cloud services, cloud databases, cloud computing, cloud functions, cloud storage, network services, cloud communications, Cloud servers for basic cloud computing services such as middleware services, domain name services, security services, CDN, and big data and artificial intelligence platforms.
- the server 120 is used to provide background services for applications supporting human-computer interaction.
- the server 120 undertakes the main model training work, and the terminal 110 undertakes the secondary model training work; or, the server 120 undertakes the secondary model training work, and the terminal 110 undertakes the main model training work; or, the server 120 or the terminal 110 may be independent Undertake model training.
- the server 120 may be composed of an access server, a model training server, and a database server.
- the access server is used to provide the terminal 110 to provide access services.
- the model training server is used for model training based on the authorized dialogue data provided by the terminal.
- the terminal 110 may generally refer to one of multiple terminals, and this embodiment only uses the terminal 110 as an example for illustration.
- the number of the aforementioned terminals may be more or less.
- the foregoing terminal may be only one, or the foregoing terminal may be tens or hundreds, or more.
- the embodiment of the method for training the dialog model also includes other terminals.
- the embodiments of the present application do not limit the number of terminals and device types.
- Fig. 2 is a flowchart of a method for training a dialogue model provided by an embodiment of the present application. This embodiment is described by taking the execution subject as a server as an example. Referring to Fig. 2, this embodiment includes:
- the server obtains at least two first dialogue features and at least two second dialogue features of the first dialogue based on the prior network and the posterior network in the dialogue model, where the prior network is used to output the probability distribution of the dialogue features,
- the posterior network is used to estimate the probability distribution of the dialogue features output by the prior network
- the first dialogue feature is used to represent the posterior features of the dialogue above and a dialogue reply in a dialogue
- the second dialogue feature is used to Represents the a priori feature of the dialogue above and a dialogue reply in a dialogue
- the first dialogue includes a first dialogue above and at least two first dialogue replies.
- step 201 the server acquires at least two first dialog features and at least two second dialog features of the first dialog, and the first dialog feature and the second dialog feature are used to represent the first dialog above and the first dialog, respectively.
- the posterior feature and the prior feature of a dialogue reply, a dialogue above corresponds to at least two dialogue replies.
- the server may select one dialogue from a plurality of dialogues as the first dialogue, and the first dialogue includes a first dialogue above and at least two first dialogue replies corresponding to the first dialogue above. For any group of the first dialogue above and the first dialogue reply, the server can obtain the corresponding prior feature and the posterior feature through the prior network and the posterior network, respectively.
- the server updates the dialogue model based on at least two first dialogue features and at least two second dialogue features of the first dialogue.
- the server may obtain at least one dialogue feature and at least two second dialogue features of the first dialogue.
- the first dialogue above and a first dialogue reply in the first dialogue can obtain a first dialogue feature and a second dialogue feature, and the dialogue model is updated once according to the first dialogue feature and the second dialogue feature , Update the parameters of the prior network and the posterior network in the dialogue model.
- another first dialogue feature and another second dialogue feature are obtained, and the dialogue model is updated again.
- the number of times the dialogue model is updated is the same as the number of first dialogue replies contained in the first dialogue.
- the dialogue model may also include an encoder, a decoder, and a discriminator. While the server updates the parameters of the a priori network and the posterior network, it will also update the parameters of the encoder, decoder, and discriminator. .
- the server updates the posterior network based on at least two first dialog features of the first dialog.
- the server may obtain at least two second dialog features of the above-mentioned first dialog, and then update the parameters of the posterior network once based on each second dialog feature.
- the server updates the discriminator of the dialog model according to the at least two first dialog features and the at least two second dialog features of the second dialog.
- the second dialogue includes one second dialogue above and at least two second dialogue replies.
- the server selects at least one dialogue from a plurality of dialogues as the second dialogue.
- the server can obtain at least two first dialogues of the second dialogue according to the method described in step 201.
- Dialogue features and at least two second dialogue features For the second dialogue above and a second dialogue reply according to any second dialogue, a first dialogue feature and a second dialogue feature can be obtained.
- the discriminator Based on the first dialogue feature and the second dialogue feature, the discriminator’s The parameters are updated once. At this time, the number of times the discriminator is updated is the number of replies to the second dialogue included in the second dialogue.
- the server can obtain the threshold of the number of iterations of the discriminator, and then perform multiple iterations according to the threshold of the number of iterations, and the training ends when the threshold of the number of iterations is reached.
- the server uses the trained model as a dialogue model.
- the training end condition may be reaching a predetermined number of iterations, or the model converges, or the output result of the model meets the target condition, or meets other training end conditions, etc.
- the embodiment of the present application does not limit this.
- the dialogue model is updated multiple times through multiple dialogue features of the first dialogue, and the posterior network is updated again, and then the discriminator of the dialogue model is updated according to the multiple dialogue characteristics of the second dialogue.
- Different semantics of the dialogue can be taken into consideration, so that the reply of the dialogue contains a variety of semantics, which improves the performance of the dialogue model and also increases the diversity of the dialogue replies generated through the dialogue model.
- Fig. 3 is a flowchart of a method for generating a dialog reply provided by an embodiment of the present application. This embodiment is described by taking the execution subject as the terminal as an example. Referring to FIG. 3, this embodiment includes:
- the terminal obtains the above conversation.
- the dialogue above may be content input by the terminal user, such as text, voice, or emoticons.
- the terminal inputs the dialogue text into the dialogue model, and based on the prior network in the dialogue model, randomly extracts a target dialogue feature from the second dialogue features corresponding to the multiple dialogue replies.
- the terminal may be provided with a dialogue model, and the content input by the user is used as the dialogue above.
- the dialogue model is used to encode the input dialogue above, and the encoded features are input into the dialogue model
- the prior network in, based on the prior network, a target dialogue feature is randomly selected from a plurality of first dialogue features. Due to the random extraction, when the terminal re-inputs the dialogue above, the dialogue feature extracted by the prior network may be different from the dialogue feature extracted last time, so the dialogue response output by the dialogue model is also different.
- the terminal decodes the target dialogue feature based on the decoder in the dialogue model, and outputs the target dialogue reply.
- the decoder in the dialogue model can decode the target dialogue features obtained by random extraction to obtain the target dialogue reply. If the dialogue features randomly extracted by the prior network are different, the dialogue replies decoded by the decoder are different.
- the terminal displays the target dialogue reply.
- the terminal may adopt a manner of voice playback, text display, or display of corresponding emoticons to display the above-mentioned target dialogue response.
- the dialogue response corresponding to the dialogue above is obtained by random extraction, so that if the dialogue model is input multiple times in the same dialogue, different dialogue responses can be obtained, thereby increasing the diversity of dialogue replies. sex.
- the terminal obtains and outputs the dialogue reply through the dialogue model configured by itself, and in some embodiments, the terminal can perform the dialogue model through the dialogue model configured on the server.
- the dialogue reply is obtained and output based on the obtained dialogue reply to achieve the effect of man-machine dialogue.
- Fig. 4 is a flowchart of a method for training a dialogue model provided by an embodiment of the present application. This embodiment is described by taking the server performing one iteration as an example. Referring to FIG. 4, this embodiment includes:
- the server obtains a first conversation from a plurality of conversations.
- the server may randomly select N dialogues from the multiple dialogues as the first dialogue, where N is a positive integer.
- the first dialogue includes a first dialogue above and K first dialogue replies corresponding to the first dialogue above, where K is a positive integer greater than or equal to 2.
- the number of first dialog replies included in different first dialogs may be the same or different.
- the data set includes 1000 conversations, and the server randomly selects 10 conversations as the first conversation, and obtains the first conversations A, B, C, D, E, F, G, H, I, and J, among which the first conversation A Corresponding to the 5 first dialogue replies a1, a2, a3, a4 and a5, the first dialogue B corresponds to the 6 first dialogue replies b1, b2, b3, b4, b5 and b6, and the first dialogue C corresponds to the 5 first dialogues Reply to c1, c2, c3, c4, c5, and c6. I won't list them all here.
- the server acquires at least two first dialog features and at least two second dialog features of the first dialog based on the a priori network and the posterior network in the dialog model, where the first dialog feature is used to represent a dialog in a dialog
- the text and the posterior feature of a dialogue reply is used to represent the prior feature of the dialogue above and a dialogue reply in a dialogue
- the first dialogue includes a first dialogue above and at least two first dialogues. Reply to a conversation.
- the server can encode the first dialogue reply and the corresponding first dialogue above, and then input the encoded vector representations into the prior network and the posterior network to obtain the prior features and Posterior features, that is, the second dialogue feature and the first dialogue feature.
- the server may obtain a pair of the first dialogue feature and the second dialogue feature based on the posterior network and the prior network, that is, A first dialogue feature and a second dialogue feature.
- the step of the server acquiring at least two first dialog features and at least two second dialog features of a first dialog can be implemented through the following sub-steps 4021 to 4023.
- the server separately encodes the first dialogue above and the first dialogue reply based on the encoder of the dialogue model to obtain the first vector above the first dialogue The second vector of the reply to the first conversation.
- the server inputs the first dialogue above and the first dialogue reply respectively into the encoder of the dialogue model, and the encoder is constructed based on a two-way gated recurrent unit neural network. According to the encoder, the server encodes the above-mentioned first dialogue and the first dialogue reply respectively to obtain the first vector of the first dialogue above and the second vector of the first dialogue reply.
- the encoder encodes all inputs, such as the first dialogue above and the first dialogue reply, through a two-way gated recurrent unit neural network, and the encoded vector is a fixed-length vector.
- the encoded vector is a fixed-length vector.
- take the first conversation above The first vector c obtained by encoding is taken as an example for description.
- the first vector c is calculated by the following formula (1) to formula (4).
- GRU() represents the gated recurrent unit, Indicates the first dialogue above The vector representation of the t-1th word from the left, Indicates the first dialogue above The code corresponding to the t-th word from the left.
- GRU() represents the gated recurrent unit, Indicates the first dialogue above The vector representation of the t+1th word from the right, Indicates the first dialogue above The code corresponding to the t-th word from the right.
- h t represents the first dialogue above The vector representation of the t-th word from the left in the middle and the first dialogue above The splicing vector represented by the vector of the t-th word from the right.
- c represents the first dialogue above
- the splicing vector represented by the vector of the first word from the right, T represents the first dialogue above The number of words included in.
- the server acquires at least two first dialogue features of the first conversation, and the first conversation feature of the first conversation is a result of the first vector above the first conversation and the reply to the first conversation through the posterior network.
- the second vector is processed.
- the server obtains the first dialogue feature based on the posterior network and according to the first vector and the second vector above the first dialogue.
- the posterior network is used to learn the distribution of the dialogue features of the dialogue based on the dialogue above and the dialogue reply. According to the reply information, the distribution of the dialogue features in the dialogue model obtained by training can be made more accurate.
- the probability distribution of the dialogue features output by the posterior network is called the posterior distribution, which is used to estimate the prior distribution, that is, the probability distribution of the dialogue features output by the prior network.
- the server is based on the posterior network, and according to the first vector and the second vector, the step of acquiring the first dialogue feature may be: the server may be based on the posterior network, according to the first vector and the first dialogue above the first dialogue
- the second vector of the reply, the mean value of the first parameter and the variance of the first parameter of the posterior distribution are obtained.
- the server may obtain the first dialog feature according to the mean value of the first parameter, the variance of the first parameter, and the first sample value.
- the first sampling value is the value obtained by sampling from the standard normal distribution, that is, the value of the sampling point.
- the decoder Since the value obtained by upsampling from the standard normal distribution is used to obtain the first dialogue feature, during the training process, the decoder reconstructs the dialogue reply based on the first dialogue feature, and based on the reconstructed dialogue reply and the first dialogue feature A difference between the dialogue replies is used to adjust the parameters of the dialogue model so that the difference between the first dialogue feature and the first dialogue reply is small, so that the first dialogue feature can be used to represent the first dialogue reply.
- the server obtains the first dialogue feature based on the posterior network, it is calculated by the following formula (5) and formula (6).
- ⁇ k represents the mean value of the first parameter
- ⁇ k represents the variance of the first parameter
- W represents the variable parameter
- g ⁇ () represents the posterior network
- x k represents the second vector of the first dialog reply
- c represents the first dialog
- the first vector above, b represents the bias parameter.
- z k represents the first dialogue feature
- ⁇ k represents the mean value of the first parameter
- ⁇ k represents the variance of the first parameter
- ⁇ represents the first sample value
- the server may obtain a second dialogue feature based on the prior network and according to the first vector and the reply category to which the first dialogue reply belongs, where the reply category includes at least one other dialogue reply belonging to the same category as the first dialogue reply .
- the prior network is used to represent the probability distribution of the real dialogue feature, which is estimated from the posterior distribution.
- at least two dialogue replies corresponding to one dialogue can be clustered to obtain multiple reply categories.
- the sub-distribution in the prior distribution is selected according to the reply category to which the reply of the first dialogue belongs.
- the server selects the sub-distribution according to the reply category to which the first dialog reply belongs, and then samples the second dialog feature from the sub-distribution.
- the server is based on a priori network, and according to the first vector and the reply category to which the first dialogue reply belongs, the step of acquiring the second dialogue feature may be: the server may according to the first vector and the reply category to which the first dialogue reply belongs , Determine the target probability distribution, the target probability distribution is the probability distribution corresponding to the reply category in the probability distribution of the dialog feature output by the prior network, that is, the sub-distribution used to match the posterior distribution.
- the server may obtain the mean value of the second parameter and the variance of the second parameter according to the first vector based on the prior network.
- the server may obtain the second dialog feature according to the second parameter mean value, the second parameter variance, and the second sample value.
- the second sampling value is the value obtained by sampling from the target probability distribution, that is, the value of the sampling point. Since the second dialogue feature is obtained by mixing the sampling values on the sub-distribution in the Gaussian distribution, in the training process, based on the encoder, the prior distribution and the posterior distribution are obtained based on the second dialogue feature and the first dialogue feature The Wasserstein distance between the distributions, so as to accurately match the prior distribution and the posterior distribution.
- At least two posterior distributions can be obtained according to the first dialogue including at least two dialogue replies, and one first dialogue feature z k can be sampled from each posterior distribution.
- a prior distribution can be obtained, the prior distribution includes at least two sub-distributions, and a second dialogue feature can be sampled from each sub-distribution That is, for the same first dialogue, at least two second dialogue features obtained From the same prior distribution.
- the server updates the dialogue model based on at least two first dialogue features and at least two second dialogue features of the first dialogue.
- the server may obtain the first dialogue feature and the second dialogue feature corresponding to the first dialogue reply.
- the server may obtain the discriminator loss and reconstruction loss according to the first vector obtained by encoding above the first dialog, the first dialog feature and the second dialog feature corresponding to the first dialog reply.
- the server can update the parameters of the posterior network and the prior network in the dialogue model according to the discriminator loss, and update the parameters of the encoder, the posterior network, the prior network and the decoder in the dialogue model according to the reconstruction loss.
- the server can update the parameters of the discriminator of the dialogue model according to the loss of the discriminator.
- the loss of the discriminator is obtained by adversarial network optimization of the Wasserstein distance between the posterior distribution and the prior distribution.
- the server is based on the discriminator of the dialogue model, and according to the first vector above the first dialogue, the first dialogue feature and the second dialogue feature corresponding to the first dialogue reply, the first dialogue feature and the second dialogue feature are obtained.
- the discriminator loss can be calculated by formula (10).
- ⁇ P-net represents the parameters of the prior network
- lr represents the learning rate of the dialogue model
- Means derivation Indicates the loss of the discriminator.
- the server updates the parameters of the posterior network in the dialogue model according to the loss of the discriminator, and the parameters of the posterior model are calculated by formula (12).
- ⁇ R-net represents the parameters of the posterior network
- lr represents the learning rate of the dialogue model
- Means derivation Indicates the loss of the discriminator.
- the reconstruction loss can be based on the first dialogue feature obtained by up-sampling the posterior distribution, and decode the first dialogue feature based on the decoder to reconstruct the dialogue reply. Based on the reconstructed dialogue reply and the first dialogue The error between the replies determines the reconstruction loss.
- the server may decode the first dialog feature based on the decoder in the dialog model, and obtain the target dialog feature corresponding to the target dialog reply obtained by decoding.
- the server may obtain the reconstruction loss according to the first vector, the first dialogue feature, the second dialogue feature, and the target dialogue feature.
- the reconstruction loss can be calculated by formula (13).
- p ⁇ () represents the decoder
- x k represents the target dialogue feature
- the server updates the parameters of the encoder, posterior network, a priori network, and decoder in the dialogue model, it can be calculated by formula (14).
- ⁇ net represents the parameters of net
- lr represents the learning rate of the dialogue model
- P-net, R-net, Dec ⁇ represents that net is one of Enc
- Enc represents encoder
- P-net represents a priori network
- R-net stands for posterior network
- Dec stands for decoder.
- the parameters of the discriminator can be calculated by formula (15).
- ⁇ Disc represents the parameters of the discriminator
- lr represents the learning rate of the dialogue model
- Means derivation Indicates the loss of the discriminator.
- the server updates the posterior network based on at least two first dialog features of the first dialog.
- the server can obtain at least two first dialog features, that is, posterior features, through the above steps.
- the server can be based on semantics.
- the optimization goal of distance is to control the semantic distance between the corresponding posterior distributions above the dialogue.
- the server may maximize the Wasserstein distance between a first dialog feature and the average value of other first dialog features by using the maximum mean difference.
- the server can update the posterior network based on the at least two first dialog features of the first dialog as follows: For any first dialog feature, the server can obtain the at least two first dialog features except for the The average value of other first dialog features except the first dialog feature, and the average value is used as the average dialog feature.
- the server may obtain the second Wasserstein distance between the first dialogue feature and the average dialogue feature, and use the second Wasserstein distance as a semantic loss.
- the server can update the parameters of the posterior network based on the semantic loss. Since the semantic distance between the posterior distribution is controlled, the prior distribution is a distinguishable multi-semantic distribution.
- the server obtains the average value of the at least two first dialog features other than the first dialog feature, it can be calculated by the following formula (16).
- K represents the number of the first dialog feature
- z i represents the i-th first dialog feature
- the server uses the following formula (17) to calculate the semantic loss.
- z k represents the first dialogue feature
- GKF() represents the Gaussian kernel function
- the mathematical expectation that the distance is large enough Mean dialogue features representing other posterior distributions The distance between is small enough for mathematical expectation.
- the parameters of the posterior network are calculated by the following formula (18).
- ⁇ R-net represents the parameters of the posterior network
- lr represents the learning rate of the dialogue model
- Means derivation represents the semantic loss.
- the server updates the discriminator of the dialogue model according to at least two first dialogue features and at least two second dialogue features of the second dialogue, where the second dialogue includes a second dialogue above and at least two second dialogue replies .
- the server can set the update times of the discriminator. Each time the discriminator is updated, the server selects at least one of the multiple conversations as the second conversation, and then obtains at least two first conversations of the second conversation For the feature and the at least two second dialog features, please refer to step 402, which will not be repeated here.
- the server can obtain the discriminator loss according to the first dialog feature and the second dialog feature corresponding to the second dialog reply. For details, please refer to step 403, which will not be repeated here.
- the server can update the discriminator in the dialogue model according to the loss of the discriminator. When the server updates the parameters of the discriminator in the dialogue model, you can refer to the above formula (15), which will not be repeated here.
- steps 401 to 405 are an iterative process of the dialog model training method provided in the embodiment of the present application, and the server repeats the above steps until the training end condition is satisfied.
- the dialogue model is updated multiple times through multiple dialogue features of the first dialogue, and the posterior network is updated again, and then the discriminator in the dialogue model is updated according to the multiple dialogue features of the second dialogue,
- the different semantics of the dialogue are considered, so that the replies of the dialogue contain a variety of semantics, and the diversity of the dialogue replies generated by the dialogue model is improved.
- FIG. 5 is a schematic structural diagram of a dialogue model provided according to an embodiment of the present application.
- a first dialogue is schematically shown on the left side, and the first dialogue includes a first dialogue above and K Reply to the first conversation.
- Inputting the first dialogue to the encoder can obtain the first vector above the first dialogue and the second vector of the first dialogue reply.
- the first vector is input into the prior network to obtain the prior distribution, and multiple second dialogue features can be sampled from each sub-distribution of the prior distribution.
- the first dialogue feature corresponding to the k-th first dialogue reply is z k , and the other first The average value of the dialogue feature is The decoder decodes the first dialogue feature z k to obtain a reconstructed dialogue reply.
- the reconstructed dialogue reply is similar to the first dialogue reply, the better.
- the following introduces the multi-semantic WAE (Wasserstein Auto-Encoder, Wasserstein Auto-Encoder) algorithm used in training the above-mentioned dialogue model in the embodiment of the present application.
- Encoder encoder
- R-net PosteriorNetwork (posterior network)
- P-net PriorNetwork (prior network); Disc: Discriminator (discriminator);
- Input Anthology The number of reply clusters K, the number of discriminator iterations n critic , and the number of model iterations max-step.
- FIG. 6 is a schematic flowchart of a multi-semantic WAE algorithm provided according to an embodiment of the present application.
- the input of the WAE algorithm is multiple dialogues.
- Step 1 is to initialize the encoder parameters;
- Step 2 is to determine the condition of the model iteration;
- Step 3 is to obtain at least one first dialogue;
- Step 4 is to respond based on the first dialogue in the first dialogue.
- Step 5 is to encode the first dialogue above and the first dialogue reply;
- Step 6 is to obtain the first dialogue feature based on the posterior network;
- Step 7 is to obtain the second dialogue feature based on the prior network;
- Step 8 To update the prior network according to the discriminator loss;
- step 9 is to update the posterior network according to the discriminator loss;
- step 10 is to update the encoder, posterior network, a priori network and decoder according to the reconstruction loss;
- step 11 is The discriminator is updated according to the discriminator loss;
- step 12 is the end of the iteration based on the first dialog reply;
- step 13 is the iterative determination based on the first dialog feature;
- step 14 is the posterior network is updated based on the semantic loss.
- Step 15 is the end of the iteration based on the first dialogue feature;
- Step 16 is the iterative determination based on the update times of the discriminator;
- Step 17 is to obtain at least one second dialogue;
- Step 18 is to perform the reply based on the second dialogue in the second dialogue Iterative judgment;
- Step 19 is to repeat the above steps 5 to 7;
- Step 20 is to update the discriminator based on the loss of the discriminator;
- Step 21 is the end of the iteration based on the second dialog reply;
- Step 22 is the number of discriminator updates plus 1;
- Step 23 is The iteration based on the update times of the discriminator ends;
- step 24 is the number of model iterations plus 1;
- step 25 is the end of model iteration.
- Dialogue data set The number to be selected M, the threshold ⁇ .
- the embodiment of the application also designs experiments for verification.
- the experiment is evaluated through two public session data sets.
- One data set is Douban (from Yu Wu, Furu Wei, Shaohan Huang, Yunli Wang, Zhoujun Li, and Ming Zhou’s "Response generation by context-aware prototype editing” published in Proceedings of the Intelligence, ArtiConference on AAAI Volume 33, pages 7281-7288).
- BLUE an automatic evaluation method of machine translation
- BOWEmbedding bagofwordsEmbedding, bag of words model embedding
- intra-dist Intrinsic difference
- inter-dist external difference
- BLUE includes Recall (recall rate), Precision (precision) and F1 (F1-Score, F1 score).
- BOWEmbedding includes Average (average value), Extreme (extreme value) and Greedy (greed value).
- Intra-dist includes dist-1 and dist-2
- inter-dist includes dist-1 and dist-2.
- Table 2 The numbers with a + sign in Table 2 indicate values that exceed the optimal basic threshold and are statistically significant.
- the data in Table 2 shows that the MA-WAE method proposed in this application significantly improves diversity and maintains relevance.
- this application also designs a manual evaluation experiment.
- 5 participants were recruited from Informativeness (informativeness, which measures whether a dialogue response provides meaningful information), Appropriateness (appropriateness, which measures whether a dialogue response is logical) and Semantic Diversity (semantic diversity) .
- the score is 0-2, with 0 being the worst and 2 being the best.
- Table 3 shows the average ⁇ standard deviation of all methods. The results show that in terms of semantic diversity, MS-WAE is much better than the other data sets on both data sets, exceeding the baseline.
- Fig. 7 is a block diagram of an apparatus for training a dialogue model provided according to an embodiment of the present application.
- the device is used to execute the steps in the execution of the above-mentioned dialog model training method.
- the device includes: a feature acquisition module 701, a model update module 702, and a model acquisition module 703.
- the feature acquisition module 701 is configured to acquire at least two first dialog features and at least two second dialog features of a first dialog, where the first dialog feature and the second dialog feature are used to represent the first dialog above and one
- the posterior feature and the prior feature of the first dialogue reply one dialogue above corresponds to at least two dialogue replies; optionally, the feature acquisition module 701 is used to acquire based on the prior network and the posterior network in the dialogue model At least two first dialogue features and at least two second dialogue features of the first dialogue, the prior network is used to output the probability distribution of the dialogue features, and the posterior network is used to estimate the value of the dialogue features output by the prior network Probability distribution, the first dialogue feature is used to represent the posterior features of the dialogue above and a dialogue reply in a dialogue, and the second dialogue feature is used to express the prior features of the dialogue above and a dialogue reply in a dialogue,
- the first dialogue includes a first dialogue above and at least two first dialogue replies.
- the model update module 702 is configured to update the dialog model based on at least two first dialog features and at least two second dialog features of the first dialog, the dialog model includes a priori network and a posterior network, the posterior network Used to estimate the probability distribution of the dialogue features output by the prior network;
- the model update module 702 is further configured to update the posterior network based on at least two first dialog features of the first dialog;
- the model update module 702 is further configured to update the discriminator of the dialog model based on at least two first dialog features and at least two second dialog features of the second dialog; the second dialog includes the second dialog above and at least Reply to two second conversations.
- the model acquisition module 703 is configured to use the trained model as a dialogue model in response to meeting the training end condition.
- the feature acquisition module 701 is configured to, for any first dialog reply, based on the dialog model, respectively encode the first dialog above and the first dialog reply to obtain the The first vector above the first dialogue and the second vector of the first dialogue reply; based on the posterior network, according to the first vector and the second vector, the first dialogue feature is obtained; based on the prior network, according to The first vector and the reply category to which the first dialogue reply belongs, and the second dialogue feature is obtained, and the reply category includes at least one other dialogue reply belonging to the same category as the first dialogue reply.
- the feature acquisition module 701 is configured to reply to any of the first dialogs of the first dialog, and based on the encoder of the dialog model, the above and the first dialog of the first dialog A dialogue reply is coded separately to obtain the first vector above the first dialogue and the second vector of the first dialogue reply; obtain at least two first dialogue features of the first dialogue, and the first dialogue of the first dialogue A dialog feature is obtained by processing the first vector above the first dialog and the second vector of the first dialog reply through the posterior network; acquiring at least two second dialog features of the first dialog, the first The second dialogue feature of the conversation is obtained by processing the first vector above the first dialogue and the reply category to which the first dialogue reply belongs through the prior network.
- the feature acquisition module 701 is configured to input the first dialogue above and the first dialogue reply into the encoder of the dialogue model respectively, and the encoder is based on a two-way gated loop unit Neural network construction; according to the encoder, the first dialogue above and the first dialogue reply are respectively encoded to obtain the first vector of the first dialogue above and the second vector of the first dialogue reply.
- the feature acquisition module 701 is further configured to acquire the mean value of the first parameter and the variance of the first parameter according to the first vector and the second vector based on the posterior network; A parameter mean, the first parameter variance, and a first sample value are used to obtain a first dialog feature, and the first sample value is a value of a sample point obtained from a standard normal distribution.
- the feature acquisition module 701 is configured to input the first vector and the second vector into the posterior network, and output the mean value of the first parameter and the variance of the first parameter; according to the first parameter
- the mean value, the variance of the first parameter, and the first sampling value are used to obtain the first dialog feature, and the first sampling value is obtained by sampling the standard normal distribution.
- the feature acquisition module 701 is configured to determine a target probability distribution according to the first vector and the reply category to which the first dialog reply belongs, and the target probability distribution is output by the prior network The probability distribution corresponding to the reply category in the probability distribution of the dialogue feature; based on the prior network, according to the first vector, obtain the second parameter mean and the second parameter variance; according to the second parameter mean and the second parameter variance And a second sampling value to obtain a second dialogue feature, where the second sampling value is a value of a sampling point obtained from the target probability distribution.
- the feature acquisition module 701 is configured to determine a target probability distribution according to the first vector and the reply category to which the first dialog reply belongs, and the target probability distribution is output by the prior network The probability distribution corresponding to the response category in the probability distribution; input the first vector into the prior network to obtain the second parameter mean and the second parameter variance; according to the second parameter mean, the second parameter variance, and the second sampling Value, the second dialog feature is obtained, and the second sampling value is obtained by sampling the probability distribution of the target.
- the model update module 702 is configured to, for any first dialogue reply of the first conversation, obtain the first dialogue feature and the second dialogue feature corresponding to the first dialogue reply;
- the first vector, the first dialog feature and the second dialog feature corresponding to the first dialog reply obtain the discriminator loss, the first vector is obtained based on the encoding of the first dialog; according to the first vector, the first dialog Reply to the corresponding first and second dialog features to obtain the reconstruction loss; according to the discriminator loss, update the parameters of the posterior network and the prior network in the dialog model; update the dialog model according to the reconstruction loss
- the parameters of the encoder, the posterior network, the prior network and the decoder; according to the discriminator loss, the parameters of the discriminator in the dialogue model are updated.
- the model update module 702 is used for a discriminator based on the dialog model, according to the first vector above the first dialog, the first dialog feature corresponding to the first dialog reply, and The second dialogue feature is to obtain the first Wasserstein distance between the first dialogue feature and the second dialogue feature corresponding to the first dialogue reply, and use the first Wassstein distance as the discriminator loss.
- the model update module 702 is configured to decode the first dialog feature based on the decoder in the dialog model to obtain the target dialog feature; according to the first vector and the first vector The first dialog feature and the second dialog feature corresponding to a dialog reply, and the target dialog feature, obtain the reconstruction loss.
- the model update module 702 is further configured to obtain, for any first dialog feature of the first dialog, the at least two first dialog features other than the first dialog feature.
- the average value of a dialogue feature the average value is used as the average dialogue feature; the second Wasserstein distance between the first dialogue feature and the average dialogue feature is obtained, and the second Wasserstein distance is taken as the semantic loss; according to This semantic loss updates the parameters of the posterior network.
- the dialog model is updated multiple times by using multiple dialog features of the first dialog, and the posterior network is updated again, and then the discriminator in the dialog model is updated based on the multiple dialog features of the second dialog.
- the parameters of the dialogue model are updated many times, taking into account the different semantics of the dialogue, so that the dialogue reply contains a variety of semantics, and the diversity of dialogue replies generated through the dialogue model is improved.
- the device for training the dialogue model provided in the above embodiment runs an application, it only uses the division of the above functional modules for illustration. In actual applications, the above functions can be allocated to different functional modules according to needs. Complete, that is, divide the internal structure of the device into different functional modules to complete all or part of the functions described above.
- the dialog model training device provided in the foregoing embodiment and the dialog model training method embodiment belong to the same concept. For the specific implementation process, please refer to the method embodiment, which will not be repeated here.
- Fig. 8 is a block diagram of a device for generating a dialog reply according to an embodiment of the present application.
- the device is used to perform the steps in the execution of the above-mentioned dialog reply generation method.
- the device includes: a dialog acquisition module 801, a feature extraction module 802, a reply output module 803, and a reply display module 804.
- the dialogue acquisition module 801 is used to acquire the dialogue above;
- the feature extraction module 802 is configured to input the dialogue text into the dialogue model, and based on the prior network in the dialogue model, randomly extract a target dialogue feature from the first dialogue features corresponding to the multiple dialogue replies;
- Reply output module 803 configured to decode the target dialogue feature based on the decoder in the dialogue model, and output the target dialogue reply;
- the reply display module 804 is used to display the target dialogue reply.
- the dialogue reply corresponding to the dialogue above is obtained by randomly extracting, so that when the same dialogue above is input multiple times, different dialogue replies can be obtained, thereby increasing the diversity of dialogue replies.
- dialog reply generating device when the dialog reply generating device provided in the above embodiment runs an application, it only uses the division of the above functional modules for illustration. In actual applications, the above functions can be allocated by different functional modules according to needs. , That is, divide the internal structure of the device into different functional modules to complete all or part of the functions described above.
- the dialog reply generating device provided in the above embodiment and the dialog reply generating method embodiment belong to the same concept. For the specific implementation process, please refer to the method embodiment, which will not be repeated here.
- the computer device can be configured as a terminal or a server.
- the terminal can be used as the executive body to implement the technical solutions provided in the embodiments of this application.
- the server can be used as the execution subject to implement the technical solutions provided in the embodiments of the present application, or the technical methods provided in the present application can be implemented through the interaction between the terminal and the server, which is not limited in the embodiments of the present application.
- FIG. 9 is a structural block diagram of a terminal 900 provided in an embodiment of the present application.
- the terminal 900 can be: a smartphone, a tablet computer, an MP3 player (Moving Picture Experts Group Audio Layer III, moving picture expert compression standard audio layer 3), MP4 (Moving Picture Experts Group Audio Layer IV, moving picture expert compressing standard audio Level 4) Player, laptop or desktop computer.
- the terminal 900 may also be called user equipment, portable terminal, laptop terminal, desktop terminal and other names.
- the terminal 900 includes a processor 901 and a memory 902.
- the processor 901 may include one or more processing cores, such as a 4-core processor, an 8-core processor, and so on.
- the processor 901 may adopt at least one hardware form among DSP (Digital Signal Processing), FPGA (Field-Programmable Gate Array), and PLA (Programmable Logic Array, Programmable Logic Array). accomplish.
- the processor 901 may also include a main processor and a coprocessor.
- the main processor is a processor used to process data in the awake state, also called a CPU (Central Processing Unit, central processing unit); the coprocessor is A low-power processor used to process data in the standby state.
- the processor 901 may be integrated with a GPU (Graphics Processing Unit, image processor), and the GPU is used for rendering and drawing content that needs to be displayed on the display screen.
- the processor 901 may also include an AI (Artificial Intelligence) processor, and the AI processor is used to process computing operations related to machine learning.
- AI Artificial Intelligence
- the memory 902 may include one or more computer-readable storage media, which may be non-transitory.
- the memory 902 may also include high-speed random access memory and non-volatile memory, such as one or more magnetic disk storage devices and flash memory storage devices.
- the non-transitory computer-readable storage medium in the memory 902 is used to store at least one instruction, and the at least one instruction is used to be executed by the processor 901 to implement the dialogue model provided in the method embodiment of the present application. Training method, or dialogue response generation method.
- the terminal 900 may optionally further include: a peripheral device interface 903 and at least one peripheral device.
- the processor 901, the memory 902, and the peripheral device interface 903 may be connected by a bus or a signal line.
- Each peripheral device can be connected to the peripheral device interface 903 through a bus, a signal line, or a circuit board.
- the peripheral device includes: at least one of a radio frequency circuit 904, a display screen 905, a camera assembly 906, an audio circuit 907, a positioning assembly 908, and a power supply 909.
- the peripheral device interface 903 can be used to connect at least one peripheral device related to I/O (Input/Output) to the processor 901 and the memory 902.
- the processor 901, the memory 902, and the peripheral device interface 903 are integrated on the same chip or circuit board; in some other embodiments, any one of the processor 901, the memory 902, and the peripheral device interface 903 or The two can be implemented on a separate chip or circuit board, which is not limited in this embodiment.
- the radio frequency circuit 904 is used for receiving and transmitting RF (Radio Frequency, radio frequency) signals, also called electromagnetic signals.
- the radio frequency circuit 904 communicates with a communication network and other communication devices through electromagnetic signals.
- the radio frequency circuit 904 converts electrical signals into electromagnetic signals for transmission, or converts received electromagnetic signals into electrical signals.
- the radio frequency circuit 904 includes: an antenna system, an RF transceiver, one or more amplifiers, a tuner, an oscillator, a digital signal processor, a codec chipset, a user identity module card, and so on.
- the radio frequency circuit 904 can communicate with other terminals through at least one wireless communication protocol.
- the wireless communication protocol includes, but is not limited to: metropolitan area networks, various generations of mobile communication networks (2G, 3G, 4G, and 5G), wireless local area networks and/or WiFi (Wireless Fidelity, wireless fidelity) networks.
- the radio frequency circuit 904 may also include a circuit related to NFC (Near Field Communication), which is not limited in this application.
- the display screen 905 is used to display a UI (User Interface, user interface).
- the UI can include graphics, text, icons, videos, and any combination thereof.
- the display screen 905 also has the ability to collect touch signals on or above the surface of the display screen 905.
- the touch signal can be input to the processor 901 as a control signal for processing.
- the display screen 905 may also be used to provide virtual buttons and/or virtual keyboards, also called soft buttons and/or soft keyboards.
- the display screen 905 may be a flexible display screen, which is disposed on the curved surface or the folding surface of the terminal 900.
- the display screen 905 can also be set as a non-rectangular irregular pattern, that is, a special-shaped screen.
- the display screen 905 may be made of materials such as LCD (Liquid Crystal Display) and OLED (Organic Light-Emitting Diode).
- the camera assembly 906 is used to capture images or videos.
- the camera assembly 906 includes a front camera and a rear camera.
- the front camera is set on the front panel of the terminal, and the rear camera is set on the back of the terminal.
- the camera assembly 906 may also include a flash.
- the flash can be a single-color flash or a dual-color flash. Dual color temperature flash refers to a combination of warm light flash and cold light flash, which can be used for light compensation under different color temperatures.
- the audio circuit 907 may include a microphone and a speaker.
- the microphone is used to collect sound waves of the user and the environment, and convert the sound waves into electrical signals and input to the processor 901 for processing, or input to the radio frequency circuit 904 to implement voice communication.
- the microphone can also be an array microphone or an omnidirectional collection microphone.
- the speaker is used to convert the electrical signal from the processor 901 or the radio frequency circuit 904 into sound waves.
- the speaker can be a traditional thin-film speaker or a piezoelectric ceramic speaker.
- the speaker When the speaker is a piezoelectric ceramic speaker, it can not only convert the electrical signal into human audible sound waves, but also convert the electrical signal into human inaudible sound waves for purposes such as distance measurement.
- the audio circuit 907 may also include a headphone jack.
- the positioning component 908 is used to locate the current geographic location of the terminal 900 to implement navigation or LBS (Location Based Service, location-based service).
- the positioning component 908 may be a positioning component based on the GPS (Global Positioning System, Global Positioning System) of the United States, the Beidou system of China, the Grenas system of Russia, or the Galileo system of the European Union.
- the power supply 909 is used to supply power to various components in the terminal 900.
- the power source 909 may be an alternating current, a direct current, a disposable battery, or a rechargeable battery.
- the rechargeable battery may support wired charging or wireless charging.
- the rechargeable battery can also be used to support fast charging technology.
- the terminal 900 further includes one or more sensors 910.
- the one or more sensors 910 include, but are not limited to: an acceleration sensor 911, a gyroscope sensor 912, a pressure sensor 913, a fingerprint sensor 914, an optical sensor 915, and a proximity sensor 916.
- the acceleration sensor 911 can detect the magnitude of acceleration on the three coordinate axes of the coordinate system established by the terminal 900.
- the acceleration sensor 911 may be used to detect the components of gravitational acceleration on three coordinate axes.
- the processor 901 may control the display screen 905 to display the user interface in a horizontal view or a vertical view according to the gravity acceleration signal collected by the acceleration sensor 911.
- the acceleration sensor 911 may also be used for the collection of game or user motion data.
- the gyroscope sensor 912 can detect the body direction and the rotation angle of the terminal 900, and the gyroscope sensor 912 can cooperate with the acceleration sensor 911 to collect the user's 3D actions on the terminal 900.
- the processor 901 can implement the following functions according to the data collected by the gyroscope sensor 912: motion sensing (for example, changing the UI according to the user's tilt operation), image stabilization during shooting, game control, and inertial navigation.
- the pressure sensor 913 may be disposed on the side frame of the terminal 900 and/or the lower layer of the display screen 905.
- the processor 901 performs left and right hand recognition or quick operation according to the holding signal collected by the pressure sensor 913.
- the processor 901 controls the operability controls on the UI interface according to the user's pressure operation on the display screen 905.
- the operability control includes at least one of a button control, a scroll bar control, an icon control, and a menu control.
- the fingerprint sensor 914 is used to collect the user's fingerprint.
- the processor 901 can identify the user's identity based on the fingerprint collected by the fingerprint sensor 914, or the fingerprint sensor 914 can identify the user's identity based on the collected fingerprints. When it is recognized that the user's identity is a trusted identity, the processor 901 authorizes the user to perform related sensitive operations, including unlocking the screen, viewing encrypted information, downloading software, paying, and changing settings.
- the fingerprint sensor 914 may be provided on the front, back or side of the terminal 900. When a physical button or a manufacturer logo is provided on the terminal 900, the fingerprint sensor 914 may be integrated with the physical button or the manufacturer logo.
- the optical sensor 915 is used to collect the ambient light intensity.
- the processor 901 may control the display brightness of the display screen 905 according to the ambient light intensity collected by the optical sensor 915. Specifically, when the ambient light intensity is high, the display brightness of the display screen 905 is increased; when the ambient light intensity is low, the display brightness of the display screen 905 is decreased.
- the processor 901 may also dynamically adjust the shooting parameters of the camera assembly 906 according to the ambient light intensity collected by the optical sensor 915.
- the proximity sensor 916 also called a distance sensor, is usually provided on the front panel of the terminal 900.
- the proximity sensor 916 is used to collect the distance between the user and the front of the terminal 900.
- the processor 901 controls the display screen 905 to switch from the on-screen state to the off-screen state; when the proximity sensor 916 detects When the distance between the user and the front of the terminal 900 gradually increases, the processor 901 controls the display screen 905 to switch from the rest screen state to the bright screen state.
- FIG. 9 does not constitute a limitation on the terminal 900, and may include more or fewer components than shown in the figure, or combine certain components, or adopt different component arrangements.
- FIG. 10 is a schematic structural diagram of a server provided according to an embodiment of the present application.
- the server 1000 may have relatively large differences due to different configurations or performance, and may include one or more processors ( Central Processing Units (CPU) 1001 and one or more memories 1002, where at least one instruction is stored in the memory 1002, and the at least one instruction is loaded and executed by the processor 1001 to implement the dialogue provided by the foregoing various method embodiments Model training method or dialogue response generation method.
- the server may also have components such as a wired or wireless network interface, a keyboard, an input and output interface for input and output, and the server 1000 may also include other components for implementing device functions, which will not be repeated here.
- the embodiment of the present application also provides a computer-readable storage medium, the computer-readable storage medium is applied to a computer device, the computer-readable storage medium stores at least one piece of program code, and the at least one piece of program code is used by a processor Execute and implement the operations performed by the computer device in the dialog model training method or the dialog reply generation method in the embodiment of the present application.
Landscapes
- Engineering & Computer Science (AREA)
- Physics & Mathematics (AREA)
- Theoretical Computer Science (AREA)
- General Physics & Mathematics (AREA)
- General Engineering & Computer Science (AREA)
- Artificial Intelligence (AREA)
- Computational Linguistics (AREA)
- Data Mining & Analysis (AREA)
- Mathematical Physics (AREA)
- General Health & Medical Sciences (AREA)
- Health & Medical Sciences (AREA)
- Databases & Information Systems (AREA)
- Evolutionary Computation (AREA)
- Computing Systems (AREA)
- Software Systems (AREA)
- Molecular Biology (AREA)
- Biophysics (AREA)
- Biomedical Technology (AREA)
- Life Sciences & Earth Sciences (AREA)
- Human Computer Interaction (AREA)
- Audiology, Speech & Language Pathology (AREA)
- Probability & Statistics with Applications (AREA)
- Algebra (AREA)
- Computational Mathematics (AREA)
- Mathematical Analysis (AREA)
- Mathematical Optimization (AREA)
- Pure & Applied Mathematics (AREA)
- Machine Translation (AREA)
- Information Transfer Between Computers (AREA)
- Information Retrieval, Db Structures And Fs Structures Therefor (AREA)
Abstract
Description
数据集 | train | valid | vest |
Douban | 894,721 | 15,000 | 15,000 |
DailyDialog | 68,096 | 6,895 | 6,695 |
Claims (14)
- 一种对话模型的训练方法,其特征在于,所述方法包括:基于对话模型中的先验网络和后验网络,获取第一对话的至少两个第一对话特征和至少两个第二对话特征,所述先验网络用于输出对话特征的概率分布,所述后验网络用于估计所述先验网络所输出的对话特征的概率分布,所述第一对话特征用于表示一个对话中对话上文和一个对话回复的后验特征,所述第二对话特征用于表示一个对话中所述对话上文和一个对话回复的先验特征,所述第一对话包括一个第一对话上文和至少两个第一对话回复;基于所述第一对话的至少两个第一对话特征和至少两个第二对话特征,更新所述对话模型;基于所述第一对话的至少两个第一对话特征,更新所述后验网络;根据第二对话的至少两个第一对话特征和至少两个第二对话特征,更新所述对话模型的判别器;响应于满足训练结束条件,将训练得到的模型作为对话模型。
- 根据权利要求1所述的方法,其特征在于,所述基于对话模型中的先验网络和后验网络,获取第一对话的至少两个第一对话特征和至少两个第二对话特征,包括:对于所述第一对话的任一所述第一对话回复,基于所述对话模型的编码器,对所述第一对话上文和所述第一对话回复分别进行编码,得到所述第一对话上文的第一向量和所述第一对话回复的第二向量;获取所述第一对话的至少两个第一对话特征,所述第一会话的所述第一对话特征通过所述后验网络对所述第一对话上文的第一向量和所述第一对话回复的第二向量进行处理得到;获取所述第一对话的至少两个第二对话特征,所述第一会话的所述第二对话特征通过所述先验网络对所述第一对话上文的第一向量和所述第一对话回复所属的回复类别进行处理得到。
- 根据权利要求2所述的方法,其特征在于,所述获取所述第一对话的至少两个第一对话特征包括:将所述第一向量和所述第二向量输入所述后验网络,输出第一参数均值和第一参数方差;根据所述第一参数均值、所述第一参数方差以及第一采样值,获取第一对话特征,所述第一采样值为对标准正态分布采样得到。
- 根据权利要求2所述的方法,其特征在于,所述获取所述第一对话的至少两个第二对话特征,包括:根据所述第一向量和所述第一对话回复所属的回复类别,确定目标概率分布,所述目标概率分布为所述先验网络所输出的概率分布中所述回复类别对应的概率分布;将所述第一向量输入所述先验网络,得到第二参数均值和第二参数方差;根据所述第二参数均值、所述第二参数方差以及第二采样值,获取第二对话特征,所述 第二采样值为对所述目标概率分布采样得到。
- 根据权利要求1所述的方法,其特征在于,所述基于所述第一对话的至少两个第一对话特征和至少两个第二对话特征,更新所述对话模型,包括:对于所述第一会话的任一所述第一对话回复,获取所述第一对话回复对应的第一对话特征和第二对话特征;根据第一向量、所述第一对话回复对应的第一对话特征和第二对话特征,获取判别器损失,所述第一向量基于所述第一对话上文编码得到;根据所述第一向量、所述第一对话回复对应的第一对话特征和第二对话特征,获取重构损失;根据所述判别器损失,更新所述对话模型中后验网络和先验网络的参数;根据所述重构损失,更新所述对话模型中编码器、所述后验网络、所述先验网络以及解码器的参数;根据所述判别器损失,更新所述对话模型的判别器的参数。
- 根据权利要求5所述的方法,其特征在于,所述根据第一向量、所述第一对话回复对应的第一对话特征和第二对话特征,获取判别器损失,包括:基于所述对话模型的判别器,根据所述第一对话上文的第一向量、所述第一对话回复对应的第一对话特征和第二对话特征,获取所述第一对话回复对应的第一对话特征和第二对话特征之间的第一瓦瑟斯坦距离,将所述第一瓦斯斯坦距离作为判别器损失。
- 根据权利要求5所述的方法,其特征在于,所述根据第一向量、所述第一对话回复对应的第一对话特征和第二对话特征,获取重构损失,包括:基于所述对话模型中的解码器,对所述第一对话特征进行解码,获取目标对话特征;根据所述第一向量、所述所述第一对话回复对应的第一对话特征和第二对话特征、所述目标对话特征,获取重构损失。
- 根据权利要求1所述的方法,其特征在于,所述基于所述第一对话的至少两个第一对话特征,更新所述所述后验网络,包括:对于所述第一对话的任一第一对话特征,获取所述至少两个第一对话特征中除所述第一对话特征外其他第一对话特征的平均值,将所述平均值作为平均对话特征;获取所述第一对话特征与所述平均对话特征之间的第二瓦瑟斯坦距离,将所述第二瓦瑟斯坦距离作为语义损失;根据所述语义损失,更新所述后验网络的参数。
- 一种对话回复生成方法,其特征在于,所述方法包括:获取对话上文;将所述对话上文输入对话模型,基于所述对话模型中的先验网络,从多个对话回复对应的第二对话特征中随机抽取一个目标对话特征;基于所述对话模型中的解码器对所述目标对话特征进行解码,输出目标对话回复;展示所述目标对话回复。
- 一种对话模型的训练装置,其特征在于,所述装置包括:特征获取模块,用于基于对话模型中的先验网络和后验网络,获取第一对话的至少两个第一对话特征和至少两个第二对话特征,所述先验网络用于输出对话特征的概率分布,所述后验网络用于估计所述先验网络所输出的对话特征的概率分布,所述第一对话特征用于表示一个对话中对话上文和一个对话回复的后验特征,所述第二对话特征用于表示一个对话中所述对话上文和一个对话回复的先验特征,所述第一对话包括一个第一对话上文和至少两个第一对话回复;模型更新模块,用于基于所述第一对话的至少两个第一对话特征和至少两个第二对话特征,更新所述对话模型;所述模型更新模块,还用于基于所述第一对话的至少两个第一对话特征,更新所述后验网络;所述模型更新模块,还用于根据第二对话的至少两个第一对话特征和至少两个第二对话特征,更新所述对话模型的判别器;模型获取模块,用于响应于满足训练结束条件,将训练得到的模型作为对话模型。
- 根据权利要求10所述的装置,其特征在于,所述特征获取模块,用于对于所述第一对话的任一所述第一对话回复,基于所述对话模型的编码器,对所述第一对话上文和所述第一对话回复分别进行编码,得到所述第一对话上文的第一向量和所述第一对话回复的第二向量;获取所述第一对话的至少两个第一对话特征,所述第一会话的所述第一对话特征通过所述后验网络对所述第一向量和所述第二向量进行处理得到;获取所述第一对话的至少两个第二对话特征,所述第一会话的所述第二对话特征通过所述先验网络对所述第一向量和所述第一对话回复所属的回复类别进行处理得到。
- 一种对话回复生成装置,其特征在于,所述装置包括:对话获取模块,用于获取对话上文;特征抽取模块,用于将所述对话上文输入对话模型,基于所述对话模型中的先验网络,从多个对话回复对应的第一对话特征中随机抽取一个目标对话特征;回复输出模块,用于基于所述对话模型中的解码器对所述目标对话特征进行解码,输出目标对话回复;回复展示模块,用于展示所述目标对话回复。
- 一种计算机设备,其特征在于,所述计算机设备包括处理器和存储器,所述存储器用于存储至少一段程序代码,所述至少一段程序代码由所述处理器加载并执行权利要求1至8所述的对话模型的训练方法,或者执行权利要求9所述的对话回复生成方法。
- 一种存储介质,其特征在于,所述存储介质用于存储至少一段程序代码,所述至少一 段程序代码用于执行权利要求1至8任一权利要求所述的对话模型的训练方法,或者执行权利要求9所述的对话回复生成方法。
Priority Applications (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
JP2022538829A JP7431977B2 (ja) | 2020-05-25 | 2021-05-06 | 対話モデルの訓練方法、装置、コンピュータ機器及びプログラム |
US17/715,778 US20220309088A1 (en) | 2020-05-25 | 2022-04-07 | Method and apparatus for training dialog model, computer device, and storage medium |
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202010450194.0A CN111680123B (zh) | 2020-05-25 | 2020-05-25 | 对话模型的训练方法、装置、计算机设备及存储介质 |
CN202010450194.0 | 2020-05-25 |
Related Child Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US17/715,778 Continuation US20220309088A1 (en) | 2020-05-25 | 2022-04-07 | Method and apparatus for training dialog model, computer device, and storage medium |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2021238599A1 true WO2021238599A1 (zh) | 2021-12-02 |
Family
ID=72434775
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/CN2021/091954 WO2021238599A1 (zh) | 2020-05-25 | 2021-05-06 | 对话模型的训练方法、装置、计算机设备及存储介质 |
Country Status (4)
Country | Link |
---|---|
US (1) | US20220309088A1 (zh) |
JP (1) | JP7431977B2 (zh) |
CN (1) | CN111680123B (zh) |
WO (1) | WO2021238599A1 (zh) |
Families Citing this family (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN111680123B (zh) * | 2020-05-25 | 2024-01-26 | 腾讯科技(深圳)有限公司 | 对话模型的训练方法、装置、计算机设备及存储介质 |
CN112702329B (zh) * | 2020-12-21 | 2023-04-07 | 四川虹微技术有限公司 | 一种流量数据异常检测方法、装置和存储介质 |
CN113254597B (zh) * | 2021-06-23 | 2021-09-28 | 腾讯科技(深圳)有限公司 | 模型训练方法、查询处理方法及相关设备 |
CN113569212B (zh) * | 2021-07-30 | 2024-04-26 | 上海交通大学 | 基于自动编码器的击键动力学身份认证与识别方法及系统 |
CN117785964A (zh) * | 2024-02-28 | 2024-03-29 | 宜宾市万事通网络信息服务有限公司 | 应用于网络服务的数据处理方法及系统 |
Citations (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US10303978B1 (en) * | 2018-03-26 | 2019-05-28 | Clinc, Inc. | Systems and methods for intelligently curating machine learning training data and improving machine learning model performance |
CN110457457A (zh) * | 2019-08-02 | 2019-11-15 | 腾讯科技(深圳)有限公司 | 对话生成模型的训练方法、对话生成方法及装置 |
CN111680123A (zh) * | 2020-05-25 | 2020-09-18 | 腾讯科技(深圳)有限公司 | 对话模型的训练方法、装置、计算机设备及存储介质 |
Family Cites Families (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
EP3454260A1 (en) * | 2017-09-11 | 2019-03-13 | Tata Consultancy Services Limited | Bilstm-siamese network based classifier for identifying target class of queries and providing responses thereof |
CN108897797A (zh) * | 2018-06-12 | 2018-11-27 | 腾讯科技(深圳)有限公司 | 对话模型的更新训练方法、装置、存储介质及电子设备 |
KR102204979B1 (ko) * | 2018-08-24 | 2021-01-19 | 네이버 주식회사 | 딥러닝 생성모델과 다중모달 분포를 이용하여 멀티턴 대화 응답을 생성하는 방법 및 시스템 |
US11544524B2 (en) * | 2018-09-28 | 2023-01-03 | Samsung Electronics Co., Ltd. | Electronic device and method of obtaining emotion information |
CN110222164B (zh) * | 2019-06-13 | 2022-11-29 | 腾讯科技(深圳)有限公司 | 一种问答模型训练方法、问题语句处理方法、装置及存储介质 |
CN111143509B (zh) * | 2019-12-09 | 2023-06-30 | 天津大学 | 一种基于静态-动态注意力变分网络的对话生成方法 |
-
2020
- 2020-05-25 CN CN202010450194.0A patent/CN111680123B/zh active Active
-
2021
- 2021-05-06 WO PCT/CN2021/091954 patent/WO2021238599A1/zh active Application Filing
- 2021-05-06 JP JP2022538829A patent/JP7431977B2/ja active Active
-
2022
- 2022-04-07 US US17/715,778 patent/US20220309088A1/en active Pending
Patent Citations (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US10303978B1 (en) * | 2018-03-26 | 2019-05-28 | Clinc, Inc. | Systems and methods for intelligently curating machine learning training data and improving machine learning model performance |
CN110457457A (zh) * | 2019-08-02 | 2019-11-15 | 腾讯科技(深圳)有限公司 | 对话生成模型的训练方法、对话生成方法及装置 |
CN111680123A (zh) * | 2020-05-25 | 2020-09-18 | 腾讯科技(深圳)有限公司 | 对话模型的训练方法、装置、计算机设备及存储介质 |
Also Published As
Publication number | Publication date |
---|---|
CN111680123B (zh) | 2024-01-26 |
CN111680123A (zh) | 2020-09-18 |
US20220309088A1 (en) | 2022-09-29 |
JP2023508062A (ja) | 2023-02-28 |
JP7431977B2 (ja) | 2024-02-15 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
WO2021238599A1 (zh) | 对话模型的训练方法、装置、计算机设备及存储介质 | |
WO2021135577A9 (zh) | 音频信号处理方法、装置、电子设备及存储介质 | |
CN110097019B (zh) | 字符识别方法、装置、计算机设备以及存储介质 | |
US20220172737A1 (en) | Speech signal processing method and speech separation method | |
CN110263131B (zh) | 回复信息生成方法、装置及存储介质 | |
CN110147533B (zh) | 编码方法、装置、设备及存储介质 | |
CN110162604B (zh) | 语句生成方法、装置、设备及存储介质 | |
CN111104980B (zh) | 确定分类结果的方法、装置、设备及存储介质 | |
CN112733970B (zh) | 图像分类模型处理方法、图像分类方法及装置 | |
CN112069309A (zh) | 信息获取方法、装置、计算机设备及存储介质 | |
CN111368525A (zh) | 信息搜索方法、装置、设备及存储介质 | |
CN111324699A (zh) | 语义匹配的方法、装置、电子设备及存储介质 | |
CN111581958A (zh) | 对话状态确定方法、装置、计算机设备及存储介质 | |
WO2020151685A1 (zh) | 编码方法、装置、设备及存储介质 | |
CN111753498A (zh) | 文本处理方法、装置、设备及存储介质 | |
CN110555102A (zh) | 媒体标题识别方法、装置及存储介质 | |
WO2022193973A1 (zh) | 图像处理方法、装置、电子设备、计算机可读存储介质及计算机程序产品 | |
CN117454954A (zh) | 模型训练方法、装置、计算机设备及存储介质 | |
CN110990549B (zh) | 获取答案的方法、装置、电子设备及存储介质 | |
CN112287070A (zh) | 词语的上下位关系确定方法、装置、计算机设备及介质 | |
CN111341307A (zh) | 语音识别方法、装置、电子设备及存储介质 | |
CN113836946B (zh) | 训练评分模型的方法、装置、终端及存储介质 | |
CN111310701B (zh) | 手势识别方法、装置、设备及存储介质 | |
CN112988984B (zh) | 特征获取方法、装置、计算机设备及存储介质 | |
CN114328815A (zh) | 文本映射模型的处理方法、装置、计算机设备及存储介质 |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 21813897 Country of ref document: EP Kind code of ref document: A1 |
|
ENP | Entry into the national phase |
Ref document number: 2022538829 Country of ref document: JP Kind code of ref document: A |
|
NENP | Non-entry into the national phase |
Ref country code: DE |
|
32PN | Ep: public notification in the ep bulletin as address of the adressee cannot be established |
Free format text: NOTING OF LOSS OF RIGHTS PURSUANT TO RULE 112(1) EPC (EPO FORM 1205A DATED 170423) |
|
122 | Ep: pct application non-entry in european phase |
Ref document number: 21813897 Country of ref document: EP Kind code of ref document: A1 |