CN112836526B - Multi-language neural machine translation method and device based on gating mechanism - Google Patents
Multi-language neural machine translation method and device based on gating mechanism Download PDFInfo
- Publication number
- CN112836526B CN112836526B CN202110132050.5A CN202110132050A CN112836526B CN 112836526 B CN112836526 B CN 112836526B CN 202110132050 A CN202110132050 A CN 202110132050A CN 112836526 B CN112836526 B CN 112836526B
- Authority
- CN
- China
- Prior art keywords
- language
- data
- target language
- word
- decoding
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Active
Links
- 238000013519 translation Methods 0.000 title claims abstract description 66
- 230000007246 mechanism Effects 0.000 title claims abstract description 44
- 238000000034 method Methods 0.000 title claims abstract description 42
- 230000001537 neural effect Effects 0.000 title claims abstract description 24
- 230000006870 function Effects 0.000 claims abstract description 20
- 238000007781 pre-processing Methods 0.000 claims abstract description 16
- 238000004364 calculation method Methods 0.000 claims abstract description 13
- 239000013598 vector Substances 0.000 claims description 58
- 239000011159 matrix material Substances 0.000 claims description 45
- 230000004927 fusion Effects 0.000 claims description 9
- 238000004422 calculation algorithm Methods 0.000 claims description 4
- 230000009466 transformation Effects 0.000 claims description 4
- 238000012545 processing Methods 0.000 claims description 3
- 230000011218 segmentation Effects 0.000 claims description 3
- 238000012549 training Methods 0.000 description 5
- 230000007547 defect Effects 0.000 description 3
- 238000012986 modification Methods 0.000 description 2
- 230000004048 modification Effects 0.000 description 2
- 238000006467 substitution reaction Methods 0.000 description 2
- 230000001360 synchronised effect Effects 0.000 description 2
- 230000009286 beneficial effect Effects 0.000 description 1
- 238000012512 characterization method Methods 0.000 description 1
- 238000004891 communication Methods 0.000 description 1
- 238000004590 computer program Methods 0.000 description 1
- 238000010586 diagram Methods 0.000 description 1
- 230000000694 effects Effects 0.000 description 1
- 238000012804 iterative process Methods 0.000 description 1
- 238000003062 neural network model Methods 0.000 description 1
- 230000008569 process Effects 0.000 description 1
- 230000003068 static effect Effects 0.000 description 1
- 238000012546 transfer Methods 0.000 description 1
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F40/00—Handling natural language data
- G06F40/40—Processing or translation of natural language
- G06F40/58—Use of machine translation, e.g. for multi-lingual retrieval, for server-side translation for client devices or for real-time translation
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F40/00—Handling natural language data
- G06F40/10—Text processing
- G06F40/12—Use of codes for handling textual entities
- G06F40/126—Character encoding
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F40/00—Handling natural language data
- G06F40/20—Natural language analysis
- G06F40/279—Recognition of textual entities
- G06F40/284—Lexical analysis, e.g. tokenisation or collocates
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F40/00—Handling natural language data
- G06F40/40—Processing or translation of natural language
- G06F40/42—Data-driven translation
- G06F40/44—Statistical methods, e.g. probability models
Abstract
The invention relates to a multilingual neural machine translation method and device based on a gating mechanism, wherein the method comprises the steps of obtaining bilingual parallel data of all language pairs, and preprocessing, wherein the bilingual parallel data comprises a source language and a target language; encoding by using the preprocessed source language to obtain encoded data; decoding the target language and the coded data by using a gating mechanism to obtain target language decoded data; and performing function calculation on the target language decoding data to obtain a target language translation result corresponding to the source language.
Description
Technical Field
The invention relates to the field of translation, in particular to a multilingual neural machine translation method and device based on a gating mechanism.
Background
Multilingual neural machine translation refers to a method of translating each source language into multiple target languages using a neural network model. The existing multilingual neural machine translation method mainly adopts an end-to-end translation method proposed by Google corporation in 2017, and the method consists of a set of encoder and decoder frameworks based on a self-attention mechanism, wherein the encoder can encode sentences in different source languages into hidden vectors which are fused with information in multiple languages, and the decoder can simultaneously generate multiple target languages by using the vectors.
The existing method uses one decoder to generate a plurality of different target languages in the decoding process, however, the grammar structure and semantic information of different languages are generally quite different, and the plurality of target languages are difficult to characterize by using one set of decoder network parameters. Therefore, in the current multi-language translation task, a phenomenon that different language words are mixed in one translation result usually occurs, and the translation effect is affected.
Disclosure of Invention
The invention provides a multi-language neural machine translation method and device based on a gating mechanism, which can solve the technical problem that different language vocabulary mixing phenomenon occurs in one translation result.
The technical scheme for solving the technical problems is as follows:
provided is a multi-language neural machine translation method based on a gating mechanism, the method comprising: obtaining bilingual parallel data of all language pairs, and preprocessing, wherein the bilingual parallel data comprises a source language and a target language; encoding by using the preprocessed source language to obtain encoded data; decoding the target language and the coded data by using a gating mechanism to obtain target language decoded data; and performing function calculation on the target language decoding data to obtain a target language translation result corresponding to the source language.
Further, the obtaining bilingual parallel data of all language pairs and preprocessing includes: performing sub-word segmentation on the training corpus in the bilingual parallel data to generate a sub-word sequence; and adding a translation direction label at the tail end of the sub-word sequence to obtain a sub-word sequence vector matrix, wherein the sub-word vector matrix comprises a source language sub-word vector matrix and a target language sub-word vector matrix.
Further, the encoding with the preprocessed source language to obtain encoded data includes: acquiring a word sequence matrix of a source language; and coding according to the word sequence matrix of the source language and the sub-word sequence vector matrix of the source language to obtain coded data.
Further, the decoding the target language and the encoded representation data by using a gating mechanism to obtain target language decoded data includes: acquiring a word sequence matrix of a target language; decoding according to the coded data and the target language subword vector matrix to obtain hidden state data of a decoder; and fusing the information to be translated and the hidden state data of the decoder by using a gating mechanism to obtain target language decoding data.
Further, the fusing the information to be translated and the hidden state data of the decoder by using a gating mechanism, and decoding the data by using the target language includes: acquiring a fusion factor coefficient and a word vector of information to be translated; and obtaining target language decoding data according to the fusion factor coefficient, the hidden state data of the decoder and the word vector of the information to be translated.
Further, performing function calculation on the target language decoding data to obtain a target language translation result corresponding to the source language, including: after carrying out linear transformation on the target language decoding data, carrying out softmax function calculation to obtain probability distribution of words in a target language vocabulary; and determining that the word corresponding to the maximum probability forms a target language translation result.
There is provided a gating mechanism-based multilingual neural machine translation device, the device comprising: the preprocessing module is used for acquiring bilingual parallel data of all language pairs and preprocessing, wherein the bilingual parallel data comprises a source language and a target language; the encoding module is used for encoding by utilizing the preprocessed source language to obtain encoded data; the decoding module is used for decoding the target language and the coded data by utilizing a gating mechanism to obtain target language decoded data; and the translation module is used for carrying out function calculation on the target language decoding data to obtain a target language translation result corresponding to the source language.
Further, the decoding module is further used for obtaining a word sequence matrix of the target language; decoding according to the coded data and the word sequence matrix of the target language to obtain hidden state data of a decoder; and fusing the information to be translated and the hidden state data of the decoder by using a gating mechanism to obtain target language decoding data.
There is provided an electronic device including: at least one processor and at least one memory; the memory is used for storing one or more program instructions; the processor is configured to execute one or more program instructions to perform a gating mechanism-based multilingual neural machine translation method.
A computer readable storage medium is provided having one or more program instructions embodied therein for performing a gating mechanism based multilingual neural machine translation method by a system.
The beneficial effects of the invention are as follows: by introducing a gating mechanism into the multilingual neural machine translation, different target languages can be distinguished in a decoding end, and the method can integrate general information and vector information of all the target languages in a decoder network, so that the method can generate a plurality of different target languages by using one set of decoder network parameters, can effectively alleviate the defect that a translation result easily contains a plurality of different language vocabularies in the current multilingual translation method, and effectively improves the quality of multilingual translation.
Drawings
FIG. 1 is a flowchart of a multi-language neural machine translation method based on a gating mechanism according to an embodiment of the present invention;
fig. 2 is a schematic structural diagram of a multilingual neural machine translation device based on a gating mechanism according to an embodiment of the present invention.
Detailed Description
The principles and features of the present invention are described below with reference to the drawings, the examples are illustrated for the purpose of illustrating the invention and are not to be construed as limiting the scope of the invention.
The embodiment of the invention provides a multilingual neural machine translation method based on a gating mechanism, and referring to fig. 1, the method comprises the following steps:
step S1: obtaining bilingual parallel data of all language pairs, and preprocessing, wherein the bilingual parallel data comprises a source language and a target language;
in this embodiment, in order to reduce the influence of the external word on the translation performance, first, the BPE method is used to segment sentences in all the training corpora to generate a subword sequence, so that the input of the encoder and the decoder networks are both the subword sequence, the input of the encoder is the subword sequence of the source language, and the input of the decoder is the subword sequence of the target language. Specifically, bilingual parallel data of all language pairs are firstly obtained, when translation is carried out, a label t representing the translation direction is added after a subword sequence of the bilingual parallel data of the language pairs, and a subword sequence vector matrix is obtained, wherein the subword sequence vector matrix comprises a source language subword vector matrix and a target language subword vector matrix. For example, when the source language is english and the target language is german, < E2D > is added, indicating that english is translated into german; when the source language is chinese and the target language is japanese, < Z2J > is added, indicating that chinese is translated into japanese. The preprocessing can lead bilingual parallel data not only to contain specific words but also to contain vectors representing the translation direction, so that the translation direction is guided, and the phenomenon of mixing multiple target languages is avoided.
The BPE (byte pair encoding) method is an algorithm for encoding according to byte pairs. The main purpose is to compress the data, and the algorithm is described as a layer-by-layer iterative process in which the most frequent pair of characters in a string is replaced by a character that does not appear in this character. The resource splits the training corpus by taking characters as a unit, combines the training corpus according to character pairs, sorts all combined results according to the frequency of occurrence, and ranks the higher the frequency of occurrence, the higher the ranking is, and the first ranking is the subword with the highest frequency of occurrence.
Step S2: encoding by using the preprocessed source language to obtain encoded data;
in this embodiment, a word sequence matrix x= [ x ] of the source language is obtained 0 ,x 1 …,x n ]The method comprises the steps of carrying out a first treatment on the surface of the With X= [ v 0 ,v 1 ,…,v n ]Representing a matrix obtained by word vector preprocessing of the input word sequence in the source language, namely a source language sub-word vector matrix, wherein v is as follows 0 Vector representing translation direction tag, v n(n>0) A vector representing the nth word. And coding according to the word sequence matrix of the source language and the sub-word sequence vector matrix of the source language to obtain coded data. Wherein Self is defined enc () For a self-attention mechanism based encoder calculation unit, the encoded representation of each word through the encoder can be calculated from the following formula:
wherein the method comprises the steps ofThe encoded representation representing the sequence of the t-th word in the n-th layer, we can get the top-most encoded representation h using the encoder N Wherein each source language shares the same set of encoder representations.
That is, the encoded data of the present embodiment isWhile the encoded data may characterize the hidden state of the encoder to facilitate decoding based on the hidden state of the encoder.
Step S3: decoding the target language and the coded data by using a gating mechanism to obtain target language decoded data;
in this embodiment, first, a word sequence matrix of a target language is obtained; decoding according to the coded data and the target language subword vector matrix to obtain hidden state data of the decoder; and finally, fusing the information to be translated and the hidden state data of the decoder by using a gating mechanism to obtain target language decoding data.
Specifically, the decoder network relies on h obtained in step S2 N And the attention mechanism module is used for obtaining word sequence decoding representation of the multi-language information. Acquiring word sequence matrix y= [ y ] of target language 1 …,y n ]And a target language subword vector matrix y= [ u ] 1 ,…,u n ]That is, a matrix obtained by word vector preprocessing of word sequences input in a target language, wherein u n A vector representing the nth word.
Further, define Self dec () For self-attention based decoder computing unit, the decoder outputs hidden state at time tCalculated from the following formula:
wherein h is N Representing the hidden state obtained by the encoder,decoding data of the nth word sequence in the nth layer of the target end, u t Representing the input of the decoder at time t, u t Comprises one target language, namely +.> At the same time (I)>Also, hidden layer unit characterization shared by all target languages.
In order to distinguish different target languages, the present embodiment adopts a gating mechanism, which can fuse the information to be translated and the hidden state data of the decoder, that is, fuse the hidden layer unit of the shared decoder and the input of one target language. And acquiring the fusion factor coefficient and the word vector of the information to be translated, and acquiring target language decoding data according to the fusion factor coefficient, the hidden state data of the decoder and the word vector of the information to be translated.
Specifically, the expression of the target language decoded data is:
wherein, the hidden state data of the decoder are fused with factor coefficient lambdaWord vector of information to be translated->Obtaining target language decoded data->The expression of the fusion factor coefficient is +.>Wherein ReLU is a linear rectification function.
Step S4: and performing function calculation on the target language decoding data to obtain a target language translation result corresponding to the source language.
In this embodiment, after performing linear transformation on the target language decoded data, performing softmax function calculation to obtain probability distribution of words in the target language vocabulary; and determining that the word corresponding to the maximum probability forms a target language translation result.
For example, the uppermost hidden state output via a decoderAfter a layer of linear transformation (), the following is shown:
wherein,is an input representation of the softmax layer of the target language i.
Obtained by linear transformationThe probability distribution of each time t in the target language vocabulary is output through a softmax function:
wherein W is i And b i Is a training parameter of the model, W i The dimension is the same as the vocabulary dimension.
Finally, selecting the word corresponding to the maximum probability as the generated result of the time t:
according to the steps, the final translation result Y= [ Y ] of the language is generated by decoding in turn 1 ,…,y n ]Wherein Y comprises L target languages.
According to the method, a gating mechanism is introduced into the multilingual neural machine translation, different target languages can be distinguished in a decoding end, and universal information and vector information of all the target languages can be fused in a decoder network, so that a set of decoder network parameters are used for generating a plurality of different target languages, the defect that a translation result easily contains a plurality of different language words in the current multilingual translation method can be effectively relieved, and the quality of multilingual translation is effectively improved.
The embodiment of the invention also provides a multilingual neural machine translation device based on a gating mechanism, and referring to fig. 2, the device comprises:
the preprocessing module 01 is used for acquiring bilingual parallel data of all language pairs and preprocessing, wherein the bilingual parallel data comprises a source language and a target language; the functions corresponding to the modules are described in the above step S1, and in order to avoid repetition, they are not described herein again.
The encoding module 02 is used for encoding by utilizing the preprocessed source language to obtain encoded data; the functions corresponding to the modules are described in the above step S2, and in order to avoid repetition, they are not described herein again.
The decoding module 03 is configured to decode the target language and the encoded data by using a gating mechanism to obtain target language decoded data; the decoding module is also used for obtaining a word sequence matrix of the target language; decoding according to the coded data and the word sequence matrix of the target language to obtain hidden state data of a decoder; and fusing the information to be translated and the hidden state data of the decoder by using a gating mechanism to obtain target language decoding data. The functions corresponding to the modules are described in the above step S3, and in order to avoid repetition, they are not described herein again.
And the translation module 04 is used for carrying out function calculation on the target language decoding data to obtain a target language translation result corresponding to the source language. The functions corresponding to the modules are described in the above step S4, and in order to avoid repetition, they are not described herein again.
According to the method, a gating mechanism is introduced into the multilingual neural machine translation, different target languages can be distinguished in a decoding end, and universal information and vector information of all the target languages can be fused in a decoder network, so that a set of decoder network parameters are used for generating a plurality of different target languages, the defect that a translation result easily contains a plurality of different language words in the current multilingual translation method can be effectively relieved, and the quality of multilingual translation is effectively improved.
The present embodiment provides an electronic device including: at least one processor and at least one memory; the memory is used for storing one or more program instructions; the processor is configured to execute one or more program instructions to perform a gating mechanism-based multilingual neural machine translation method.
The present embodiment provides a computer readable storage medium having one or more program instructions embodied therein for performing a gating mechanism-based multilingual neural machine translation method by a system.
In the embodiment of the invention, the processor may be an integrated circuit chip with signal processing capability. The processor may be a general purpose processor, a digital signal processor (Digital Signal Processor, DSP for short), an application specific integrated circuit (Application Specific f ntegrated Circuit ASIC for short), a field programmable gate array (FieldProgrammable Gate Array FPGA for short), or other programmable logic device, discrete gate or transistor logic device, discrete hardware components.
The disclosed methods, steps, and logic blocks in the embodiments of the present invention may be implemented or performed. A general purpose processor may be a microprocessor or the processor may be any conventional processor or the like. The steps of the method disclosed in connection with the embodiments of the present invention may be embodied directly in the execution of a hardware decoding processor, or in the execution of a combination of hardware and software modules in a decoding processor. The software modules may be located in a random access memory, flash memory, read only memory, programmable read only memory, or electrically erasable programmable memory, registers, etc. as well known in the art. The processor reads the information in the storage medium and, in combination with its hardware, performs the steps of the above method.
The storage medium may be memory, for example, may be volatile memory or nonvolatile memory, or may include both volatile and nonvolatile memory.
The nonvolatile Memory may be a Read-Only Memory (ROM), a Programmable ROM (PROM), an Erasable PROM (EPROM), an electrically Erasable ROM (Electrically EPROM, EEPROM), or a flash Memory.
The volatile memory may be a random access memory (Random Access Memory, RAM for short) which acts as an external cache. By way of example, and not limitation, many forms of RAM are available, such as Static RAM (SRAM), dynamic RAM (DRAM), synchronous DRAM (SDRAM), double data rate SDRAM (Double Data RateSDRAM), enhanced SDRAM (ESDRAM), synchronous DRAM (SLDRAM), and direct memory bus RAM (directracram, DRRAM).
The storage media described in embodiments of the present invention are intended to comprise, without being limited to, these and any other suitable types of memory.
Those skilled in the art will appreciate that in one or more of the examples described above, the functions described in the present invention may be implemented in a combination of hardware and software. When the software is applied, the corresponding functions may be stored in a computer-readable medium or transmitted as one or more instructions or code on the computer-readable medium. Computer-readable media includes both computer storage media and communication media including any medium that facilitates transfer of a computer program from one place to another. A storage media may be any available media that can be accessed by a general purpose or special purpose computer.
The present invention is not limited to the above embodiments, and various equivalent modifications and substitutions can be easily made by those skilled in the art within the technical scope of the present invention, and these modifications and substitutions are intended to be included in the scope of the present invention. Therefore, the protection scope of the invention is subject to the protection scope of the claims.
Claims (6)
1. A method for multilingual neural machine translation based on a gating mechanism, the method comprising:
obtaining bilingual parallel data of all language pairs, and preprocessing, wherein the bilingual parallel data comprises a source language and a target language;
encoding by using the preprocessed source language to obtain encoded data;
decoding the target language and the coded data by using a gating mechanism to obtain target language decoded data;
performing function calculation on the target language decoding data to obtain a target language translation result corresponding to the source language;
wherein the decoding the target language and the encoded data by using the gating mechanism to obtain target language decoded data includes: decoding according to the coded data and the target language subword vector matrix to obtain hidden state data of a decoder; fusing the information to be translated and the hidden state data of the decoder by using a gating mechanism to obtain target language decoding data;
the method for decoding data by using the gating mechanism to fuse the information to be translated and the hidden state data of the decoder comprises the following steps: acquiring a fusion factor coefficient and a word vector of information to be translated; according to the fusion factor coefficient, the hidden state data of the decoder and the word vector of the information to be translated, utilizing a gating mechanism to fuse, and obtaining target language decoding data;
the obtaining bilingual parallel data of all language pairs and preprocessing comprises the following steps:
performing subword segmentation on the bilingual parallel data by using a BPE algorithm to generate a subword sequence;
adding a translation direction label at the tail end of a sub-word sequence of bilingual parallel data of a language pair, and obtaining a sub-word sequence vector matrix through word vector processing, wherein the sub-word sequence vector matrix comprises a source language sub-word vector matrix and a target language sub-word vector matrix, and X= [ v ] is used 0 ,v 1 ,…,v n ]Representing a source language subword vector matrix, where v 0 Vector representing translation direction tag, v n The vector representing the nth word is represented by y= [ u ] 1 ,…,u n ]Representing a target language subword vector matrix, where u n A vector representing the nth word.
2. The method for multiple language neural machine translation based on gating mechanism of claim 1, wherein the encoding with the preprocessed source language to obtain encoded data comprises:
acquiring a word sequence matrix of a source language;
and coding according to the word sequence matrix of the source language and the sub-word sequence vector matrix of the source language to obtain coded data.
3. The method for multi-language neural machine translation based on a gating mechanism according to claim 1, wherein performing a function calculation on the target language decoded data to obtain a target language translation result corresponding to a source language comprises:
after carrying out linear transformation on the target language decoding data, carrying out softmax function calculation to obtain probability distribution of words in a target language vocabulary;
and determining that the word corresponding to the maximum probability forms a target language translation result.
4. A gating mechanism-based multilingual neural machine translation device, the device comprising:
the preprocessing module is used for acquiring bilingual parallel data of all language pairs and preprocessing, wherein the bilingual parallel data comprises a source language and a target language;
the encoding module is used for encoding by utilizing the preprocessed source language to obtain encoded data;
the decoding module is used for decoding the target language and the coded data by utilizing a gating mechanism to obtain target language decoded data;
the translation module is used for carrying out function calculation on the target language decoding data to obtain a target language translation result corresponding to the source language;
wherein, the decoding module is further configured to perform the following operations: decoding according to the coded data and the target language subword vector matrix to obtain hidden state data of a decoder; acquiring a fusion factor coefficient and a word vector of information to be translated; according to the fusion factor coefficient, the hidden state data of the decoder and the word vector of the information to be translated, utilizing a gating mechanism to fuse, and obtaining target language decoding data;
the preprocessing module is further used for performing subword segmentation on the bilingual parallel data by using a BPE algorithm to generate a subword sequence;
adding a translation direction label at the tail end of a sub-word sequence of bilingual parallel data of a language pair, and obtaining a sub-word sequence vector matrix through word vector processing, wherein the sub-word sequence vector matrix is formed byThe word sequence vector matrix comprises a source language subword vector matrix and a target language subword vector matrix, and X= [ v ] 0 ,v 1 ,…,v n ]Representing a source language subword vector matrix, where v 0 Vector representing translation direction tag, v n The vector representing the nth word is represented by y= [ u ] 1 ,…,u n ]Representing a target language subword vector matrix, where u n A vector representing the nth word.
5. An electronic device, the electronic device comprising: at least one processor and at least one memory;
the memory is used for storing one or more program instructions;
the processor being operative to execute one or more program instructions for performing the method as claimed in any one of claims 1-3.
6. A computer readable storage medium, characterized in that the computer storage medium contains one or more program instructions for performing the method according to any of claims 1-3 by a system.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202110132050.5A CN112836526B (en) | 2021-01-31 | 2021-01-31 | Multi-language neural machine translation method and device based on gating mechanism |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202110132050.5A CN112836526B (en) | 2021-01-31 | 2021-01-31 | Multi-language neural machine translation method and device based on gating mechanism |
Publications (2)
Publication Number | Publication Date |
---|---|
CN112836526A CN112836526A (en) | 2021-05-25 |
CN112836526B true CN112836526B (en) | 2024-01-30 |
Family
ID=75932451
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN202110132050.5A Active CN112836526B (en) | 2021-01-31 | 2021-01-31 | Multi-language neural machine translation method and device based on gating mechanism |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN112836526B (en) |
Families Citing this family (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN115130479A (en) * | 2022-04-13 | 2022-09-30 | 腾讯科技(深圳)有限公司 | Machine translation method, target translation model training method, and related program and device |
Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN109299479A (en) * | 2018-08-21 | 2019-02-01 | 苏州大学 | Translation memory is incorporated to the method for neural machine translation by door control mechanism |
CN109543824A (en) * | 2018-11-30 | 2019-03-29 | 腾讯科技(深圳)有限公司 | A kind for the treatment of method and apparatus of series model |
CN110674646A (en) * | 2019-09-06 | 2020-01-10 | 内蒙古工业大学 | Mongolian Chinese machine translation system based on byte pair encoding technology |
CN110781690A (en) * | 2019-10-31 | 2020-02-11 | 北京理工大学 | Fusion and compression method of multi-source neural machine translation model |
Family Cites Families (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US10867595B2 (en) * | 2017-05-19 | 2020-12-15 | Baidu Usa Llc | Cold fusing sequence-to-sequence models with language models |
-
2021
- 2021-01-31 CN CN202110132050.5A patent/CN112836526B/en active Active
Patent Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN109299479A (en) * | 2018-08-21 | 2019-02-01 | 苏州大学 | Translation memory is incorporated to the method for neural machine translation by door control mechanism |
CN109543824A (en) * | 2018-11-30 | 2019-03-29 | 腾讯科技(深圳)有限公司 | A kind for the treatment of method and apparatus of series model |
CN110674646A (en) * | 2019-09-06 | 2020-01-10 | 内蒙古工业大学 | Mongolian Chinese machine translation system based on byte pair encoding technology |
CN110781690A (en) * | 2019-10-31 | 2020-02-11 | 北京理工大学 | Fusion and compression method of multi-source neural machine translation model |
Non-Patent Citations (1)
Title |
---|
基于数据扩充的翻译记忆库与神经机器翻译融合方法;曹骞;熊德意;;中文信息学报(第05期);40-47 * |
Also Published As
Publication number | Publication date |
---|---|
CN112836526A (en) | 2021-05-25 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN108416058B (en) | Bi-LSTM input information enhancement-based relation extraction method | |
CN111247581B (en) | Multi-language text voice synthesizing method, device, equipment and storage medium | |
CN105068998A (en) | Translation method and translation device based on neural network model | |
KR102254612B1 (en) | method and device for retelling text, server and storage medium | |
CN112699690B (en) | Translation model training method, translation method, electronic device and storage medium | |
CN112800757B (en) | Keyword generation method, device, equipment and medium | |
JP2021033995A (en) | Text processing apparatus, method, device, and computer-readable storage medium | |
CN113536795B (en) | Method, system, electronic device and storage medium for entity relation extraction | |
CN113590784A (en) | Triple information extraction method and device, electronic equipment and storage medium | |
CN112016271A (en) | Language style conversion model training method, text processing method and device | |
CN112732902A (en) | Cross-language abstract generation method and device, electronic equipment and computer readable medium | |
CN112836526B (en) | Multi-language neural machine translation method and device based on gating mechanism | |
CN111563380A (en) | Named entity identification method and device | |
CN115759062A (en) | Knowledge injection-based text and image pre-training model processing method and text and image retrieval system | |
CN110913229B (en) | RNN-based decoder hidden state determination method, device and storage medium | |
CN112836040B (en) | Method and device for generating multilingual abstract, electronic equipment and computer readable medium | |
CN112926344B (en) | Word vector replacement data enhancement-based machine translation model training method and device, electronic equipment and storage medium | |
CN113449081A (en) | Text feature extraction method and device, computer equipment and storage medium | |
CN112052329A (en) | Text abstract generation method and device, computer equipment and readable storage medium | |
Chimalamarri et al. | Linguistically enhanced word segmentation for better neural machine translation of low resource agglutinative languages | |
CN113704466B (en) | Text multi-label classification method and device based on iterative network and electronic equipment | |
US20220083745A1 (en) | Method, apparatus and electronic device for determining word representation vector | |
CN109325110B (en) | Indonesia document abstract generation method and device, storage medium and terminal equipment | |
CN115577680B (en) | Ancient book text sentence-breaking method and device and ancient book text sentence-breaking model training method | |
Watve et al. | English to hindi translation using transformer |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
GR01 | Patent grant | ||
GR01 | Patent grant |