CN117273027A - Automatic machine translation post-verification method based on translation error correction - Google Patents
Automatic machine translation post-verification method based on translation error correction Download PDFInfo
- Publication number
- CN117273027A CN117273027A CN202311559241.5A CN202311559241A CN117273027A CN 117273027 A CN117273027 A CN 117273027A CN 202311559241 A CN202311559241 A CN 202311559241A CN 117273027 A CN117273027 A CN 117273027A
- Authority
- CN
- China
- Prior art keywords
- translation
- machine
- machine translation
- sequence
- error
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Granted
Links
- 238000013519 translation Methods 0.000 title claims abstract description 229
- 238000012795 verification Methods 0.000 title claims abstract description 58
- 238000000034 method Methods 0.000 title claims abstract description 43
- 238000012937 correction Methods 0.000 title claims abstract description 24
- 238000012549 training Methods 0.000 claims abstract description 39
- 238000013528 artificial neural network Methods 0.000 claims abstract description 20
- 238000006243 chemical reaction Methods 0.000 claims abstract description 15
- 238000003062 neural network model Methods 0.000 claims abstract description 13
- 230000014616 translation Effects 0.000 claims description 205
- 238000002372 labelling Methods 0.000 claims description 10
- 230000001323 posttranslational effect Effects 0.000 claims description 6
- 230000004913 activation Effects 0.000 claims description 3
- 238000004364 calculation method Methods 0.000 claims description 3
- 239000011159 matrix material Substances 0.000 claims description 3
- 230000007246 mechanism Effects 0.000 claims description 3
- 238000005457 optimization Methods 0.000 claims description 3
- 210000005036 nerve Anatomy 0.000 claims description 2
- 238000012545 processing Methods 0.000 description 9
- 238000004590 computer program Methods 0.000 description 8
- 230000006870 function Effects 0.000 description 7
- 238000010586 diagram Methods 0.000 description 3
- 238000012986 modification Methods 0.000 description 3
- 230000004048 modification Effects 0.000 description 3
- 230000008569 process Effects 0.000 description 3
- 238000003491 array Methods 0.000 description 2
- 230000009286 beneficial effect Effects 0.000 description 2
- 238000013461 design Methods 0.000 description 2
- 230000003287 optical effect Effects 0.000 description 2
- 238000013527 convolutional neural network Methods 0.000 description 1
- 238000009472 formulation Methods 0.000 description 1
- 239000000203 mixture Substances 0.000 description 1
- 238000011160 research Methods 0.000 description 1
- 230000009182 swimming Effects 0.000 description 1
- 238000012360 testing method Methods 0.000 description 1
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F40/00—Handling natural language data
- G06F40/40—Processing or translation of natural language
- G06F40/58—Use of machine translation, e.g. for multi-lingual retrieval, for server-side translation for client devices or for real-time translation
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/21—Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
- G06F18/214—Generating training patterns; Bootstrap methods, e.g. bagging or boosting
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F40/00—Handling natural language data
- G06F40/20—Natural language analysis
- G06F40/253—Grammatical analysis; Style critique
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F40/00—Handling natural language data
- G06F40/30—Semantic analysis
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
- G06N3/048—Activation functions
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Physics & Mathematics (AREA)
- General Engineering & Computer Science (AREA)
- General Physics & Mathematics (AREA)
- Artificial Intelligence (AREA)
- Computational Linguistics (AREA)
- General Health & Medical Sciences (AREA)
- Health & Medical Sciences (AREA)
- Audiology, Speech & Language Pathology (AREA)
- Data Mining & Analysis (AREA)
- Evolutionary Computation (AREA)
- Life Sciences & Earth Sciences (AREA)
- Bioinformatics & Cheminformatics (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Evolutionary Biology (AREA)
- Bioinformatics & Computational Biology (AREA)
- Biomedical Technology (AREA)
- Biophysics (AREA)
- Molecular Biology (AREA)
- Computing Systems (AREA)
- Mathematical Physics (AREA)
- Software Systems (AREA)
- Machine Translation (AREA)
Abstract
The invention belongs to the technical field of machine translation, and provides a machine translation automatic post-verification method based on translation error correction, wherein the method comprises the following steps: (1) Training a neural network model of the dual-to-edit sequence by adopting fine-granularity machine translated verification data as training data to obtain a neural network conversion model of the dual-to-edit sequence; (2) Based on the neural network conversion model of the dual-to-edit sequence obtained in the step (1), taking a translation original text S and a machine translation text MT as inputs, and outputting a machine translation error sequence E corresponding to the machine translation text; (3) According to the machine translation error sequence E, the machine translation MT is converted into a target translation. According to the invention, the fine-granularity translation error type is introduced as additional information, so that on one hand, the sparsity of the training data of the input end can be improved, and on the other hand, compared with a text, the sparse training data is shorter and the spatial distribution is denser, thereby relieving the dependence of the model on the large-scale training data.
Description
Technical Field
The invention belongs to the technical field of machine translation, and particularly relates to a machine translation automatic post-verification method based on translation error correction.
Background
Automatic machine translation post-verification (Machine Translation Automatic Post Editing, MTAPE for short) has received a growing interest in the research field in recent years. Generally, the automatic post-verification system for machine translation takes the translation result of the machine translation system as input, automatically corrects errors in translation, and outputs accurate and smooth translations which are more in line with the habit of human writing. The existing MTAPE system takes a triplet formed by (original text, machine translation and post-verification translation) as training data, performs parameter updating of a neural network through sentence-level end-to-end training, and obtains an MTAPE model after training convergence. In practical application, the MTAPE system and the machine translation system are isolated from each other, and the output of the machine translation system is optimized in a pipeline mode, so that a translation text with higher quality is finally obtained.
Although having the advantages of simple data form and easy training, the prior MTAPE system is limited by a sentence-to-sentence training mode, and has the following disadvantages:
(1) The training data of the triples takes sentences as basic units, so that the training data is very sparse, and the model is required to be trained by depending on a large amount of data, which is not practical in a real scene;
(1) The automatic post-verification of machine translation usually does not carry out continuous translation modification in a large range, but adjusts in a plurality of small ranges, so that the existing sentence level training mode can cause that a large amount of training signals are wasted in nonsensical translation copies by the model, and the learning difficulty of the model is increased;
(3) The existing end-to-end verification mode exists in a black box mode, and a human cannot understand the internal judgment basis of the system, so that when an error occurs in the automatic machine translation post-verification system, the system accuracy cannot be improved by aiming at the error design correction scheme.
Disclosure of Invention
The invention aims to provide a machine translation automatic post-verification method based on translation error correction, which aims to solve the technical problems existing in the prior art.
In order to achieve the above purpose, the technical scheme adopted by the invention is as follows:
a machine translation automatic post-verification method based on translation error correction, comprising:
(1) Training a neural network model of the dual-to-edit sequence by adopting fine-granularity machine translated verification data as training data to obtain a neural network conversion model of the dual-to-edit sequence;
(2) Based on the neural network conversion model of the dual-to-edit sequence obtained in the step (1) to translate the original textAnd machine translation text->As input, output of the machine translation error sequence corresponding to the machine translation text +.>;
(3) According to the sequence of machine turning errorsTranslation of text->Converting into a target translation;
the fine-granularity machine translated verification data is obtained by manually marking translation errors of phrase levels on a machine translation result.
In one embodiment, the fine-grained machine post-translational verification data is synthesized by manual annotationMachine translation error sequence obtained from three sequences +.>,/>Wherein->Represents the nth machine turning error, +.>Representing the length of the machine-turning error sequence;
which represents the translation context; />Which represents the machine translation;which represents a post-verification translation, wherein +.>Represents the first, j,/in translation, machine translation and post-verification translation, respectively>Personal word (s)/(s)>Respectively represent->The length of the three sequences.
In one embodiment, the machine translation error sequenceIs represented by a quadruple: mechanically turning the phrase sequence, the modified sequence, the original text corresponding sequence and the error type; the error types include mistranslations, missed translations, transitional translations, professional vocabulary errors, grammar errors, and/or fluency errors.
In one embodiment, the neural network model of (1) coupled to the edit sequence comprises an textual encoderMachine translation, &>And decoder->The method comprises the steps of carrying out a first treatment on the surface of the Wherein, original text encoder->Machine translation, &>For coding translation originals, respectively>And machine translation text->Decoder->With original encoder->Machine translation, &>Takes as input the output hidden state of (2) to generate the machine translation error sequence in autoregressive manner>。
In one embodiment, the textual encoderMachine translation, &>For coding translation originals, respectively>And machine translation text->Converting it into a corresponding hidden state vector:
wherein->Is composed of m layers of neural network coding layers, wherein +.>Is composed of n nerve network coding layers.
In one implementation, the decoderIn the state of translation original text hidden->Hidden state of machine translation>And the already generated machine translation error sequence +.>Predicting the machine translation error sequence at the current instant t>:
Wherein W is word vector matrix of decoder output layer,>is formed by stacking k decoding layers and one output layer.
In one embodiment, the decoding layer includes three sub-networks: context means network, textual attention network and machine translation textual mixed attention network;
the context represents that the network updates the current state according to the generated vector;
the textual attention network uses a multi-headed attention mechanism to update the hidden state vector at the current time:wherein->Representing the state of translation origin in hidden->Representing a hidden state vector at the current moment;
the machine translation text mixed attention network updates the hidden state vector of the current moment with the original text attention networkAn organic translation text hidden state->For input, updated state vector is outputWherein->Is a mixed coefficient for measuring the importance of a machine translation to a decoding model, said +.>The calculation mode of (2) is as follows:
wherein->Representing a sigmoid activation function, W and b are model parameters, +.>Is the initial mixing coefficient, d represents the input state vector +.>Is a dimension of (c).
In one embodiment, the (1) further comprises a given instance input (S, MT, E), constructing an optimized loss function for model training:
wherein (1)>For the initial character +.>For translating the length of the error sequence E, +.>For translating textual hidden state vector, +.>For machine translation hidden state vector, +.>Is an already generated machine translation error sequence.
In one embodiment, the neural network conversion model formulation for the dual-to-edit sequence in (1) is formulated as follows:
wherein (1)>Representing the set of all training set instances, which uses the fine-grained machine post-translational verification data, +.>Representing translation originals,/->Representing machine translation->Representing a machine translation error sequence,/->Is a model parameter.
In one embodiment, the specific method of (3) is as follows: according to the sequence of machine turning errorsThe error type, the position information and the text editing method based on the error type, and the automatic translation restoration is realized by correcting the machine translation through deleting and/or inserting text editing actions.
In order to achieve the above object, the present invention further provides a machine translation automatic post-verification system, including:
the model building module is used for training a neural network model of the dual-to-edit sequence by adopting fine-granularity machine translated verification data as training data to obtain a neural network conversion model of the dual-to-edit sequence;
the machine-turning error prediction module is used for translating the original text based on a neural network conversion model of a dual-to-edit sequenceAnd machine translation text->As input, the machine corresponding to the translation text of the output machine turns overTranslation error sequence +.>;
Automatic machine translation text restoration module based on machine turning error sequenceTranslation of text->Converting into a target translation;
the fine-granularity machine translated verification data is obtained by manually marking translation errors of phrase levels on a machine translation result.
To achieve the above object, the present invention also provides a computer-readable storage medium having stored thereon a computer program to be executed by a processor to implement the machine translation automatic post-verification method based on translation error correction as described above.
In order to achieve the above object, the present invention further provides a machine translation automatic post-verification device, including: a processor and a memory;
the memory is used for storing a computer program;
the processor is connected to the memory and is configured to execute the computer program stored in the memory, so that the automatic post-machine-translation checking device performs the automatic post-machine-translation checking method based on the correction of the translation error as described above.
Compared with the prior art, the invention has the following beneficial effects:
according to the invention, the fine-granularity translation error type is introduced as additional information, and the supervision signal is divided into phrase levels from sentence levels by the fine-granularity translation error label, so that the sparsity of the training data of the input end can be improved; on the other hand, the invention uses a translation error sequence as output, the sequence is shorter than the text, and the space distribution is more dense, thereby relieving the dependence of the model on large-scale training data.
The invention adopts the translation error labeling, so that the model can pay more attention to the sequences with errors, and the interference of the sequences without errors on the model is reduced, thereby reducing the learning difficulty of the model.
The invention outputs the checked translation and the translation error of the corresponding segment, is beneficial to human understanding of the internal judgment basis of the machine translation, and can construct training data according to the corresponding error type to perform model optimization, thereby improving automatic post-check of the machine translation.
Drawings
Fig. 1 is a schematic diagram of the principle of the present invention-example 1.
Fig. 2 is a schematic diagram of a neural network model coupled to an edit sequence in example 1 of the present invention.
FIG. 3 is a schematic diagram of the system of the present invention, which is a machine translation automatic post-verification based on translation error correction in embodiment 2.
Fig. 4 is a schematic structural view of the device disclosed in the embodiment 4 of the present invention.
Description of the embodiments
The present invention will be further described in detail with reference to examples so as to enable those skilled in the art to more clearly understand and understand the present invention. It should be understood that the following specific embodiments are only for explaining the present invention, and it is convenient to understand that the technical solutions provided by the present invention are not limited to the technical solutions provided by the following embodiments, and the technical solutions provided by the embodiments should not limit the protection scope of the present invention.
It should be noted that the illustrations provided in the following embodiments merely illustrate the basic concept of the present invention by way of illustration, so that only the components related to the present invention are shown in the drawings and are not drawn according to the number, shape and size of the components in actual implementation, the form, number and proportion of each component in actual implementation may be arbitrarily changed, and the layout of the components may be more complicated.
Example 1
As shown in fig. 1 and 2, this embodiment provides a machine translation automatic post-verification method based on translation error correction, which is different from the existing whole sentence verification method and adopts a fine-granularity labeling corpusTraining of a machine post-translational test model, in particular: firstly, performing phrase-level translation error labeling on a machine-turning result to obtain fine-granularity machine-translated check corpus, in the model training process, predicting a machine-turning error sequence instead of directly predicting a target translation by a model, and finally, converting the machine-translated text into the target translation according to the machine-turning error sequence, wherein the specific steps are as follows: 1. training a neural network model of the dual-to-edit sequence by adopting fine-granularity machine translated verification data as training data to obtain a neural network conversion model of the dual-to-edit sequence; 2. neural network conversion model based on obtained pair-to-edit sequence to translate original textAnd machine translation text->As input, outputting machine translation error sequence corresponding to machine translation textThe method comprises the steps of carrying out a first treatment on the surface of the 3. According to the sequence of machine turning errors->Translation of text->And converting into target translation.
The fine-grained machine post-translation verification data is synthesized by manual labelingMachine translation error sequence obtained from three sequences +.>Wherein (1)>Represents the nth machine turning error, +.>Representing the length of the machine-turning error sequence; />Which represents the translation context; />Which represents the machine translation;which represents a post-verification translation, wherein +.>Representing translation origin, machine translation and post-verification translation respectively +.>Personal word (s)/(s)>Respectively represent->The length of the three sequences.
In this embodiment, the error sequence is machine translatedIs represented by a quadruple:types. As shown in Table one, error types include misinterpretation, miss-interpretation, transitional interpretation, professional vocabulary errors, grammar errors, and/or fluency errors; phrases include the translation, definition, and/or context of the phrase; the foregoing "including" means "including but not limited to"; for example: given the translation textual input "do you like to swim in winter? "is the machine translation system output" Do you like to run in the windy winter "and the machine translated verification target" Do you like to swim in winter.
Table one: machine translation error type
In a further preferred scheme, in the process of labeling the machine-turning error sequence, the embodiment adopts a left-to-right labeling scheme, namely, preferentially labeling the machine-turning sequence with the starting point being far to the left; for different potential labeling sequences, the embodiment preferentially selects the shortest one for labeling.
In a further preferred embodiment, to avoid the problem of overlong output sequences due to phrases, the error sequence is machine translatedThe machine-turned phrase sequence and the original text corresponding sequence of (a) are each further represented as a binary group +.>Wherein->The start and end subscripts of the sequence are indicated, respectively. For example:<run, swim, swimming, transliteration>The run and swim in (2) will be replaced by +.>。
The neural network model coupled to the edit sequence in this embodiment includes an original encoderAnd decoder->The method comprises the steps of carrying out a first treatment on the surface of the Wherein, original text encoder->For coding translation originals, respectively>And machine translation text->Decoder->With original encoder->Takes as input the output hidden state of (2) to generate the machine translation error sequence in autoregressive manner>。
Specifically, given a source language inputAnd machine translation output->Original encoder->For coding translation originals, respectively>And machine translation text->Converting it into a corresponding hidden state vector:
wherein->Is composed of m layers of neural network coding layers, wherein +.>The method comprises n layers of neural network coding layers; a plurality of sequence coding models such as Transformer, CNN, RNN can be used.
DecoderIn the state of translation original text hidden-> And the already generated machine translation error sequence +.>For input, the machine translation error sequence +.>:
Wherein W is word vector matrix of decoder output layer,>the decoding layer is formed by stacking k decoding layers and one output layer, and each decoding layer comprises three sub-networks: the context represents a network, an original text attention network and a machine translation text mixed attention network, and the working principle of each sub-network is as follows:
(1) The context means that the network updates the current state according to the generated vector;
(2) The textual Attention network uses a Multi-Head Attention mechanism (Multi-Head Attention) to update the hidden state vector at the current time:wherein->Representing the state of translation origin in hidden->Representing a hidden state vector at the current moment;
(3) Machine translation of hidden state vector at current moment of mixed attention network updated with original attention networkAn organic translation text hidden state->For input, updated state vector is outputWherein->Is a mixed coefficient for measuring the importance degree of the machine translation on the decoding model, and the mixed coefficient can be used for relieving the translation error problem brought by a machine translation system, namely +.>The specific calculation mode of (2) is as follows: />Wherein->Representing a sigmoid activation function, W and b are model parameters, +.>Is the initial mixing coefficient, d represents the input state vector +.>Is a dimension of (c).
To sum up, given the instance inputs (S, MT, E), an optimized loss function for model training is constructed:
wherein (1)>For the initial character +.>For translating the length of the error sequence E, +.>For translating textual hidden state vector, +.>For machine translation hidden state vector, +.>The machine translation error sequence that has been generated.
Training the neural network model of the dual-to-edit sequence according to the constructed optimized loss function, training the neural network model in small batches, and training the neural network model by using an optimizer (such as Adam, SGD and the like) based on random gradient descent.
After model training, a neural network conversion model formula of the dual-to-edit sequence is obtained as follows:
wherein (1)>Representing a set of all training set instances, which employs fine-grained machine post-translational verification data (whereby training data may be constructed for model optimization based on corresponding error types),representing translation originals,/->Representing machine translation->Representing a machine translation error sequence,/->Is a model parameter.
After model training, the original text is translatedAnd machine translation text->As input, output of the machine translation error sequence corresponding to the machine translation text +.>(machine turning error sequence->Each term is represented by a quadruple:type), finally, according to the machine-turning error sequence +.>The error type, the position information and the text editing method based on the error type, and the automatic translation restoration is realized by correcting the machine translation through deleting and/or inserting text editing actions.
As shown in table two, the text editing method based on the error type is as follows: if the error type is mistranslation, excessive translation, grammar error, professional vocabulary error and fluency error, deleting the wrong translation and inserting the correct translation; if the error type is miss-translated, then the correct translation is inserted.
And (II) table: error type and corresponding editing operation
It should be noted that in the actual verification process (based on the real data), some error often needs to be modified multiple times, and in this case, the operations of "delete" and "insert" need to be called multiple times to correct the translation.
Example 2
As shown in fig. 3, the present embodiment provides a machine translation automatic post-verification system, which includes:
the model building module is used for training a neural network model of the dual-to-edit sequence by adopting fine-granularity machine translated verification data as training data to obtain a neural network conversion model of the dual-to-edit sequence; the fine-granularity machine translated verification data is obtained by manually marking translation errors of phrase levels on a machine translation result;
the machine-turning error prediction module is used for translating the original text based on a neural network conversion model of a dual-to-edit sequenceAnd machine translation text->As input, output of the machine translation error sequence corresponding to the machine translation text +.>;
Automatic machine translation text restoration module based on machine turning error sequenceTranslation of text->Converting into target translation, wherein the machine-turned error sequence +.>Each term is represented by a quadruple: />Types.
It should be noted that the structure and/or principle of each module corresponds to the content of the automatic post-verification method for machine translation based on translation error correction described in embodiment 1, and thus will not be described herein.
It should be noted that, it should be understood that the division of each module of the above system is merely a division of a logic function, and may be fully or partially integrated into a physical entity in actual implementation, or may be physically separated, and the modules may be fully implemented in a form of software called by a processing element; or can be realized in hardware; the method can also be realized in a form of calling software by a processing element, and the method can be realized in a form of hardware by a part of modules. For example, a module may be a processing element that is set up separately, may be implemented in a chip of an apparatus, may be stored in a memory of the apparatus in the form of program codes, may be called by a processing element of the apparatus and perform functions of a module, and may be implemented similarly. In addition, all or part of the modules can be integrated together or can be independently implemented. The processing element described herein may be an integrated circuit having signal processing capabilities. In implementation, each step of the above method or each module above may be implemented by an integrated logic circuit of hardware in a processor element or an instruction in a software form.
For example, the modules above may be one or more integrated circuits configured to implement the methods above, such as: one or more specific integrated circuits, or one or more microprocessors, or one or more field programmable gate arrays, etc. For another example, when a module above is implemented in the form of a processing element scheduler code, the processing element may be a general purpose processor, such as a central processing unit or other processor that may invoke the program code. For another example, the modules may be integrated together and implemented in a system-on-chip form.
Example 3
The present embodiment provides a computer-readable storage medium having stored thereon a computer program that is executed by a processor to implement the machine translation automatic post-verification method based on translation error correction provided in embodiment 1. Those skilled in the art will appreciate that: all or part of the steps of implementing the method provided in embodiment 1 may be implemented by hardware associated with a computer program, where the computer program may be stored in a computer readable storage medium, and when executed, the program performs steps including the method provided in embodiment 1; and the storage medium includes: various media that can store program code, such as ROM, RAM, magnetic or optical disks.
Example 4
As shown in fig. 4, this embodiment provides a machine translation automatic post-verification device, including: a processor and a memory; the memory is used for storing a computer program; the processor is connected to the memory, and is configured to execute the computer program stored in the memory, so that the automatic post-machine-translation verification device performs the automatic post-machine-translation verification method based on translation error correction described in embodiment 1.
Specifically, the memory includes: various media capable of storing program codes, such as ROM, RAM, magnetic disk, U-disk, memory card, or optical disk.
Preferably, the processor may be a general-purpose processor, including a central processor, a network processor, etc.; but also digital signal processors, application specific integrated circuits, field programmable gate arrays or other programmable logic devices, discrete gate or transistor logic devices, discrete hardware components.
The foregoing is a preferred embodiment of the present invention. It should be noted that those skilled in the art may make several modifications without departing from the design principles and technical solutions of the present invention, and these modifications should also be considered as the protection scope of the present invention.
Claims (10)
1. A machine translation automatic post-verification method based on translation error correction, comprising:
(1) Training a neural network model of the dual-to-edit sequence by adopting fine-granularity machine translated verification data as training data to obtain a neural network conversion model of the dual-to-edit sequence;
(2) Based on the neural network conversion model of the dual-to-edit sequence obtained in the step (1) to translate the original textAnd machine translation text->As input, inputMachine translation error sequence corresponding to outgoing translation text +.>;
(3) According to the sequence of machine turning errorsTranslation of text->Converting into a target translation;
the fine-granularity machine translated verification data is obtained by manually marking translation errors of phrase levels on a machine translation result.
2. The automatic post-verification method for machine translation based on translation error correction according to claim 1, wherein: the fine-grained machine post-translation verification data is synthesized by manual labelingMachine translation error sequence obtained from three sequences +.>,/>Wherein->Represents the nth machine turning error, +.>Representing the length of the machine-turning error sequence;
which represents the translation context; />Which represents the machine translation;which represents a post-verification translation, wherein +.>Representing translation origin, machine translation and post-verification translation respectively +.>Personal word (s)/(s)>Respectively represent->The length of the three sequences.
3. The automatic post-verification method for machine translation based on translation error correction according to claim 2, wherein: the machine translation error sequenceIs represented by a quadruple: mechanically turning the phrase sequence, the modified sequence, the original text corresponding sequence and the error type; the error types include mistranslations, missed translations, transitional translations, professional vocabulary errors, grammar errors, and/or fluency errors.
4. A machine translation automatic post-verification method based on translation error correction according to claim 3, wherein: the neural network model of the dual-to-edit sequence in the step (1) comprises an original text encoderMachine translation->And decoder->The method comprises the steps of carrying out a first treatment on the surface of the Wherein, original text encoder->Machine translation->For coding translation originals, respectively>And machine translation text->Decoder->With original encoder->Machine translation->Takes as input the output hidden state of (2) to generate the machine translation error sequence in autoregressive manner>。
5. The automatic post-translational error correction based machine translation verification method of claim 4, wherein said textual encoderMachine translation->For coding translation originals, respectively>And machine translation text->Converting it into a corresponding hidden state vector:
wherein->Is composed of m layers of neural network coding layers, wherein +.>Is composed of n nerve network coding layers.
6. A machine translation automatic post-verification method based on translation error correction according to claim 5, wherein said decoderIn the state of translation original text hidden->Hidden state of machine translation>And the already generated machine translation error sequence +.>For input, the machine translation error sequence +.> Wherein W is word vector matrix of decoder output layer,>is formed by stacking k decoding layers and one output layer.
7. The automatic post-translation verification based on translation error correction method according to claim 6, wherein said decoding layer comprises three sub-networks: context means network, textual attention network and machine translation textual mixed attention network;
the context represents that the network updates the current state according to the generated vector;
the textual attention network uses a multi-headed attention mechanism to update the hidden state vector at the current time:wherein->Show the state of translation original text hidden->Representing a hidden state vector at the current moment;
the machine translation text mixed attention network updates the hidden state vector of the current moment with the original text attention networkAn organic translation text hidden state->For input, updated state vector is outputWherein->Is a mixed coefficient for measuring the importance of a machine translation to a decoding model, said +.>The calculation mode of (2) is as follows:
wherein->Representing a sigmoid activation function, W and b are model parameters, +.>Initial mixing coefficient, d represents the input state vector +.>Is a dimension of (c).
8. The automatic post-verification of machine translation based on translation error correction of claim 7, wherein (1) further comprises a given instance input (S, MT, E), constructing a model trained optimization loss function:wherein (1)>For the initial character +.>For translating the length of the error sequence E, +.>For translating textual hidden state vector, +.>For machine translation hidden state vector, +.>Is an already generated machine translation error sequence.
9. The automatic post-verification method for machine translation based on translation error correction according to claim 8, wherein the neural network conversion model formula of the dual-to-edit sequence in (1) is expressed as follows:wherein (1)>Representing the set of all training set instances, which uses the fine-grained machine post-translational verification data, +.>Representing translation originals,/->Representing machine translation->Representing a machine translation error sequence,/->Is a model parameter.
10. The automatic post-verification method for machine translation based on translation error correction according to claim 9, wherein the specific method of (3) is as follows: according to the sequence of machine turning errorsThe error type, the position information and the text editing method based on the error type, and the automatic translation restoration is realized by correcting the machine translation through deleting and/or inserting text editing actions.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202311559241.5A CN117273027B (en) | 2023-11-22 | 2023-11-22 | Automatic machine translation post-verification method based on translation error correction |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202311559241.5A CN117273027B (en) | 2023-11-22 | 2023-11-22 | Automatic machine translation post-verification method based on translation error correction |
Publications (2)
Publication Number | Publication Date |
---|---|
CN117273027A true CN117273027A (en) | 2023-12-22 |
CN117273027B CN117273027B (en) | 2024-04-30 |
Family
ID=89220006
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN202311559241.5A Active CN117273027B (en) | 2023-11-22 | 2023-11-22 | Automatic machine translation post-verification method based on translation error correction |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN117273027B (en) |
Citations (11)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20190129946A1 (en) * | 2017-10-30 | 2019-05-02 | Sdl Inc. | Fragment Recall and Adaptive Automated Translation |
CN110069790A (en) * | 2019-05-10 | 2019-07-30 | 东北大学 | It is a kind of by translation retroversion to machine translation system and method literally |
CN111597778A (en) * | 2020-04-15 | 2020-08-28 | 哈尔滨工业大学 | Method and system for automatically optimizing machine translation based on self-supervision |
CN112287696A (en) * | 2020-10-29 | 2021-01-29 | 语联网(武汉)信息技术有限公司 | Post-translation editing method and device, electronic equipment and storage medium |
CN112347795A (en) * | 2020-10-04 | 2021-02-09 | 北京交通大学 | Machine translation quality evaluation method, device, equipment and medium |
CN113408307A (en) * | 2021-07-14 | 2021-09-17 | 北京理工大学 | Neural machine translation method based on translation template |
CN115345181A (en) * | 2022-07-04 | 2022-11-15 | 中国科学院自动化研究所 | Training method of neural machine translation model, translation method and device |
CN116502653A (en) * | 2022-01-18 | 2023-07-28 | 华为技术有限公司 | Translation quality evaluation method, electronic device, chip and readable storage medium |
CN116822534A (en) * | 2023-07-06 | 2023-09-29 | 四川语言桥信息技术有限公司 | Fine granularity characteristic-based machine turning evaluation index interpretation method, interpreter model and computer readable storage medium |
CN116956946A (en) * | 2023-07-14 | 2023-10-27 | 上海一者信息科技有限公司 | Machine translation text fine granularity error type identification and positioning method |
CN116992892A (en) * | 2023-07-06 | 2023-11-03 | 四川语言桥信息技术有限公司 | Method, system and readable storage medium for improving APE model based on data enhancement and multitasking training |
-
2023
- 2023-11-22 CN CN202311559241.5A patent/CN117273027B/en active Active
Patent Citations (11)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20190129946A1 (en) * | 2017-10-30 | 2019-05-02 | Sdl Inc. | Fragment Recall and Adaptive Automated Translation |
CN110069790A (en) * | 2019-05-10 | 2019-07-30 | 东北大学 | It is a kind of by translation retroversion to machine translation system and method literally |
CN111597778A (en) * | 2020-04-15 | 2020-08-28 | 哈尔滨工业大学 | Method and system for automatically optimizing machine translation based on self-supervision |
CN112347795A (en) * | 2020-10-04 | 2021-02-09 | 北京交通大学 | Machine translation quality evaluation method, device, equipment and medium |
CN112287696A (en) * | 2020-10-29 | 2021-01-29 | 语联网(武汉)信息技术有限公司 | Post-translation editing method and device, electronic equipment and storage medium |
CN113408307A (en) * | 2021-07-14 | 2021-09-17 | 北京理工大学 | Neural machine translation method based on translation template |
CN116502653A (en) * | 2022-01-18 | 2023-07-28 | 华为技术有限公司 | Translation quality evaluation method, electronic device, chip and readable storage medium |
CN115345181A (en) * | 2022-07-04 | 2022-11-15 | 中国科学院自动化研究所 | Training method of neural machine translation model, translation method and device |
CN116822534A (en) * | 2023-07-06 | 2023-09-29 | 四川语言桥信息技术有限公司 | Fine granularity characteristic-based machine turning evaluation index interpretation method, interpreter model and computer readable storage medium |
CN116992892A (en) * | 2023-07-06 | 2023-11-03 | 四川语言桥信息技术有限公司 | Method, system and readable storage medium for improving APE model based on data enhancement and multitasking training |
CN116956946A (en) * | 2023-07-14 | 2023-10-27 | 上海一者信息科技有限公司 | Machine translation text fine granularity error type identification and positioning method |
Non-Patent Citations (2)
Title |
---|
YAFU LI 等: "On compositional generalization of neural machine translation", 《COMPUTATION AND LANGUAGE》, pages 1 - 14 * |
周纯: "神经网络机器翻译技术背景下新闻文本译后编辑研究", 《品位经典》, no. 4, pages 29 - 30 * |
Also Published As
Publication number | Publication date |
---|---|
CN117273027B (en) | 2024-04-30 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN109446534B (en) | Machine translation method and device | |
CN110069790B (en) | Machine translation system and method for contrasting original text through translated text retranslation | |
CN108170686B (en) | Text translation method and device | |
CN107273356B (en) | Artificial intelligence based word segmentation method, device, server and storage medium | |
CN113811946A (en) | End-to-end automatic speech recognition of digital sequences | |
CN110532573A (en) | A kind of interpretation method and system | |
CN108132932B (en) | Neural machine translation method with replication mechanism | |
CN109522403A (en) | A kind of summary texts generation method based on fusion coding | |
WO2021127817A1 (en) | Speech synthesis method, device, and apparatus for multilingual text, and storage medium | |
WO2021082427A1 (en) | Rhythm-controlled poem generation method and apparatus, and device and storage medium | |
CN111177348B (en) | Training method and device for problem generation model, electronic equipment and storage medium | |
WO2022088570A1 (en) | Method and apparatus for post-editing of translation, electronic device, and storage medium | |
CN111144140B (en) | Zhongtai bilingual corpus generation method and device based on zero-order learning | |
CN110807335A (en) | Translation method, device, equipment and storage medium based on machine learning | |
CN110222350A (en) | By bilingual predefined translation to the method for incorporating neural Machine Translation Model | |
CN114925170B (en) | Text proofreading model training method and device and computing equipment | |
US11615247B1 (en) | Labeling method and apparatus for named entity recognition of legal instrument | |
CN115906815A (en) | Error correction method and device for modifying one or more types of wrong sentences | |
CN116822464A (en) | Text error correction method, system, equipment and storage medium | |
CN113626563A (en) | Method and electronic equipment for training natural language processing model and natural language processing | |
CN117195922A (en) | Human-in-loop neural machine translation method, system and readable storage medium | |
Zhao et al. | Tibetan multi-dialect speech recognition using latent regression Bayesian network and end-to-end mode | |
CN117273027B (en) | Automatic machine translation post-verification method based on translation error correction | |
CN114330375A (en) | Term translation method and system based on fixed paradigm | |
CN115374784A (en) | Chinese named entity recognition method based on multi-mode information selective fusion |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
GR01 | Patent grant | ||
GR01 | Patent grant |