CN117273027A - Automatic machine translation post-verification method based on translation error correction - Google Patents

Automatic machine translation post-verification method based on translation error correction Download PDF

Info

Publication number
CN117273027A
CN117273027A CN202311559241.5A CN202311559241A CN117273027A CN 117273027 A CN117273027 A CN 117273027A CN 202311559241 A CN202311559241 A CN 202311559241A CN 117273027 A CN117273027 A CN 117273027A
Authority
CN
China
Prior art keywords
translation
machine
machine translation
sequence
error
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN202311559241.5A
Other languages
Chinese (zh)
Other versions
CN117273027B (en
Inventor
白雪峰
朱宪超
张岳
霍展羽
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Sichuan Lan Bridge Information Technology Co ltd
Westlake University
Original Assignee
Sichuan Lan Bridge Information Technology Co ltd
Westlake University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Sichuan Lan Bridge Information Technology Co ltd, Westlake University filed Critical Sichuan Lan Bridge Information Technology Co ltd
Priority to CN202311559241.5A priority Critical patent/CN117273027B/en
Publication of CN117273027A publication Critical patent/CN117273027A/en
Application granted granted Critical
Publication of CN117273027B publication Critical patent/CN117273027B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F40/00Handling natural language data
    • G06F40/40Processing or translation of natural language
    • G06F40/58Use of machine translation, e.g. for multi-lingual retrieval, for server-side translation for client devices or for real-time translation
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/21Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
    • G06F18/214Generating training patterns; Bootstrap methods, e.g. bagging or boosting
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F40/00Handling natural language data
    • G06F40/20Natural language analysis
    • G06F40/253Grammatical analysis; Style critique
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F40/00Handling natural language data
    • G06F40/30Semantic analysis
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/048Activation functions

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Engineering & Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Artificial Intelligence (AREA)
  • Computational Linguistics (AREA)
  • General Health & Medical Sciences (AREA)
  • Health & Medical Sciences (AREA)
  • Audiology, Speech & Language Pathology (AREA)
  • Data Mining & Analysis (AREA)
  • Evolutionary Computation (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Evolutionary Biology (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • Biomedical Technology (AREA)
  • Biophysics (AREA)
  • Molecular Biology (AREA)
  • Computing Systems (AREA)
  • Mathematical Physics (AREA)
  • Software Systems (AREA)
  • Machine Translation (AREA)

Abstract

The invention belongs to the technical field of machine translation, and provides a machine translation automatic post-verification method based on translation error correction, wherein the method comprises the following steps: (1) Training a neural network model of the dual-to-edit sequence by adopting fine-granularity machine translated verification data as training data to obtain a neural network conversion model of the dual-to-edit sequence; (2) Based on the neural network conversion model of the dual-to-edit sequence obtained in the step (1), taking a translation original text S and a machine translation text MT as inputs, and outputting a machine translation error sequence E corresponding to the machine translation text; (3) According to the machine translation error sequence E, the machine translation MT is converted into a target translation. According to the invention, the fine-granularity translation error type is introduced as additional information, so that on one hand, the sparsity of the training data of the input end can be improved, and on the other hand, compared with a text, the sparse training data is shorter and the spatial distribution is denser, thereby relieving the dependence of the model on the large-scale training data.

Description

Automatic machine translation post-verification method based on translation error correction
Technical Field
The invention belongs to the technical field of machine translation, and particularly relates to a machine translation automatic post-verification method based on translation error correction.
Background
Automatic machine translation post-verification (Machine Translation Automatic Post Editing, MTAPE for short) has received a growing interest in the research field in recent years. Generally, the automatic post-verification system for machine translation takes the translation result of the machine translation system as input, automatically corrects errors in translation, and outputs accurate and smooth translations which are more in line with the habit of human writing. The existing MTAPE system takes a triplet formed by (original text, machine translation and post-verification translation) as training data, performs parameter updating of a neural network through sentence-level end-to-end training, and obtains an MTAPE model after training convergence. In practical application, the MTAPE system and the machine translation system are isolated from each other, and the output of the machine translation system is optimized in a pipeline mode, so that a translation text with higher quality is finally obtained.
Although having the advantages of simple data form and easy training, the prior MTAPE system is limited by a sentence-to-sentence training mode, and has the following disadvantages:
(1) The training data of the triples takes sentences as basic units, so that the training data is very sparse, and the model is required to be trained by depending on a large amount of data, which is not practical in a real scene;
(1) The automatic post-verification of machine translation usually does not carry out continuous translation modification in a large range, but adjusts in a plurality of small ranges, so that the existing sentence level training mode can cause that a large amount of training signals are wasted in nonsensical translation copies by the model, and the learning difficulty of the model is increased;
(3) The existing end-to-end verification mode exists in a black box mode, and a human cannot understand the internal judgment basis of the system, so that when an error occurs in the automatic machine translation post-verification system, the system accuracy cannot be improved by aiming at the error design correction scheme.
Disclosure of Invention
The invention aims to provide a machine translation automatic post-verification method based on translation error correction, which aims to solve the technical problems existing in the prior art.
In order to achieve the above purpose, the technical scheme adopted by the invention is as follows:
a machine translation automatic post-verification method based on translation error correction, comprising:
(1) Training a neural network model of the dual-to-edit sequence by adopting fine-granularity machine translated verification data as training data to obtain a neural network conversion model of the dual-to-edit sequence;
(2) Based on the neural network conversion model of the dual-to-edit sequence obtained in the step (1) to translate the original textAnd machine translation text->As input, output of the machine translation error sequence corresponding to the machine translation text +.>
(3) According to the sequence of machine turning errorsTranslation of text->Converting into a target translation;
the fine-granularity machine translated verification data is obtained by manually marking translation errors of phrase levels on a machine translation result.
In one embodiment, the fine-grained machine post-translational verification data is synthesized by manual annotationMachine translation error sequence obtained from three sequences +.>,/>Wherein->Represents the nth machine turning error, +.>Representing the length of the machine-turning error sequence;
which represents the translation context; />Which represents the machine translation;which represents a post-verification translation, wherein +.>Represents the first, j,/in translation, machine translation and post-verification translation, respectively>Personal word (s)/(s)>Respectively represent->The length of the three sequences.
In one embodiment, the machine translation error sequenceIs represented by a quadruple: mechanically turning the phrase sequence, the modified sequence, the original text corresponding sequence and the error type; the error types include mistranslations, missed translations, transitional translations, professional vocabulary errors, grammar errors, and/or fluency errors.
In one embodiment, the neural network model of (1) coupled to the edit sequence comprises an textual encoderMachine translation, &>And decoder->The method comprises the steps of carrying out a first treatment on the surface of the Wherein, original text encoder->Machine translation, &>For coding translation originals, respectively>And machine translation text->Decoder->With original encoder->Machine translation, &>Takes as input the output hidden state of (2) to generate the machine translation error sequence in autoregressive manner>
In one embodiment, the textual encoderMachine translation, &>For coding translation originals, respectively>And machine translation text->Converting it into a corresponding hidden state vector:
wherein->Is composed of m layers of neural network coding layers, wherein +.>Is composed of n nerve network coding layers.
In one implementation, the decoderIn the state of translation original text hidden->Hidden state of machine translation>And the already generated machine translation error sequence +.>Predicting the machine translation error sequence at the current instant t>
Wherein W is word vector matrix of decoder output layer,>is formed by stacking k decoding layers and one output layer.
In one embodiment, the decoding layer includes three sub-networks: context means network, textual attention network and machine translation textual mixed attention network;
the context represents that the network updates the current state according to the generated vector;
the textual attention network uses a multi-headed attention mechanism to update the hidden state vector at the current time:wherein->Representing the state of translation origin in hidden->Representing a hidden state vector at the current moment;
the machine translation text mixed attention network updates the hidden state vector of the current moment with the original text attention networkAn organic translation text hidden state->For input, updated state vector is outputWherein->Is a mixed coefficient for measuring the importance of a machine translation to a decoding model, said +.>The calculation mode of (2) is as follows:
wherein->Representing a sigmoid activation function, W and b are model parameters, +.>Is the initial mixing coefficient, d represents the input state vector +.>Is a dimension of (c).
In one embodiment, the (1) further comprises a given instance input (S, MT, E), constructing an optimized loss function for model training:
wherein (1)>For the initial character +.>For translating the length of the error sequence E, +.>For translating textual hidden state vector, +.>For machine translation hidden state vector, +.>Is an already generated machine translation error sequence.
In one embodiment, the neural network conversion model formulation for the dual-to-edit sequence in (1) is formulated as follows:
wherein (1)>Representing the set of all training set instances, which uses the fine-grained machine post-translational verification data, +.>Representing translation originals,/->Representing machine translation->Representing a machine translation error sequence,/->Is a model parameter.
In one embodiment, the specific method of (3) is as follows: according to the sequence of machine turning errorsThe error type, the position information and the text editing method based on the error type, and the automatic translation restoration is realized by correcting the machine translation through deleting and/or inserting text editing actions.
In order to achieve the above object, the present invention further provides a machine translation automatic post-verification system, including:
the model building module is used for training a neural network model of the dual-to-edit sequence by adopting fine-granularity machine translated verification data as training data to obtain a neural network conversion model of the dual-to-edit sequence;
the machine-turning error prediction module is used for translating the original text based on a neural network conversion model of a dual-to-edit sequenceAnd machine translation text->As input, the machine corresponding to the translation text of the output machine turns overTranslation error sequence +.>
Automatic machine translation text restoration module based on machine turning error sequenceTranslation of text->Converting into a target translation;
the fine-granularity machine translated verification data is obtained by manually marking translation errors of phrase levels on a machine translation result.
To achieve the above object, the present invention also provides a computer-readable storage medium having stored thereon a computer program to be executed by a processor to implement the machine translation automatic post-verification method based on translation error correction as described above.
In order to achieve the above object, the present invention further provides a machine translation automatic post-verification device, including: a processor and a memory;
the memory is used for storing a computer program;
the processor is connected to the memory and is configured to execute the computer program stored in the memory, so that the automatic post-machine-translation checking device performs the automatic post-machine-translation checking method based on the correction of the translation error as described above.
Compared with the prior art, the invention has the following beneficial effects:
according to the invention, the fine-granularity translation error type is introduced as additional information, and the supervision signal is divided into phrase levels from sentence levels by the fine-granularity translation error label, so that the sparsity of the training data of the input end can be improved; on the other hand, the invention uses a translation error sequence as output, the sequence is shorter than the text, and the space distribution is more dense, thereby relieving the dependence of the model on large-scale training data.
The invention adopts the translation error labeling, so that the model can pay more attention to the sequences with errors, and the interference of the sequences without errors on the model is reduced, thereby reducing the learning difficulty of the model.
The invention outputs the checked translation and the translation error of the corresponding segment, is beneficial to human understanding of the internal judgment basis of the machine translation, and can construct training data according to the corresponding error type to perform model optimization, thereby improving automatic post-check of the machine translation.
Drawings
Fig. 1 is a schematic diagram of the principle of the present invention-example 1.
Fig. 2 is a schematic diagram of a neural network model coupled to an edit sequence in example 1 of the present invention.
FIG. 3 is a schematic diagram of the system of the present invention, which is a machine translation automatic post-verification based on translation error correction in embodiment 2.
Fig. 4 is a schematic structural view of the device disclosed in the embodiment 4 of the present invention.
Description of the embodiments
The present invention will be further described in detail with reference to examples so as to enable those skilled in the art to more clearly understand and understand the present invention. It should be understood that the following specific embodiments are only for explaining the present invention, and it is convenient to understand that the technical solutions provided by the present invention are not limited to the technical solutions provided by the following embodiments, and the technical solutions provided by the embodiments should not limit the protection scope of the present invention.
It should be noted that the illustrations provided in the following embodiments merely illustrate the basic concept of the present invention by way of illustration, so that only the components related to the present invention are shown in the drawings and are not drawn according to the number, shape and size of the components in actual implementation, the form, number and proportion of each component in actual implementation may be arbitrarily changed, and the layout of the components may be more complicated.
Example 1
As shown in fig. 1 and 2, this embodiment provides a machine translation automatic post-verification method based on translation error correction, which is different from the existing whole sentence verification method and adopts a fine-granularity labeling corpusTraining of a machine post-translational test model, in particular: firstly, performing phrase-level translation error labeling on a machine-turning result to obtain fine-granularity machine-translated check corpus, in the model training process, predicting a machine-turning error sequence instead of directly predicting a target translation by a model, and finally, converting the machine-translated text into the target translation according to the machine-turning error sequence, wherein the specific steps are as follows: 1. training a neural network model of the dual-to-edit sequence by adopting fine-granularity machine translated verification data as training data to obtain a neural network conversion model of the dual-to-edit sequence; 2. neural network conversion model based on obtained pair-to-edit sequence to translate original textAnd machine translation text->As input, outputting machine translation error sequence corresponding to machine translation textThe method comprises the steps of carrying out a first treatment on the surface of the 3. According to the sequence of machine turning errors->Translation of text->And converting into target translation.
The fine-grained machine post-translation verification data is synthesized by manual labelingMachine translation error sequence obtained from three sequences +.>Wherein (1)>Represents the nth machine turning error, +.>Representing the length of the machine-turning error sequence; />Which represents the translation context; />Which represents the machine translation;which represents a post-verification translation, wherein +.>Representing translation origin, machine translation and post-verification translation respectively +.>Personal word (s)/(s)>Respectively represent->The length of the three sequences.
In this embodiment, the error sequence is machine translatedIs represented by a quadruple:types. As shown in Table one, error types include misinterpretation, miss-interpretation, transitional interpretation, professional vocabulary errors, grammar errors, and/or fluency errors; phrases include the translation, definition, and/or context of the phrase; the foregoing "including" means "including but not limited to"; for example: given the translation textual input "do you like to swim in winter? "is the machine translation system output" Do you like to run in the windy winter "and the machine translated verification target" Do you like to swim in winter.
Table one: machine translation error type
In a further preferred scheme, in the process of labeling the machine-turning error sequence, the embodiment adopts a left-to-right labeling scheme, namely, preferentially labeling the machine-turning sequence with the starting point being far to the left; for different potential labeling sequences, the embodiment preferentially selects the shortest one for labeling.
In a further preferred embodiment, to avoid the problem of overlong output sequences due to phrases, the error sequence is machine translatedThe machine-turned phrase sequence and the original text corresponding sequence of (a) are each further represented as a binary group +.>Wherein->The start and end subscripts of the sequence are indicated, respectively. For example:<run, swim, swimming, transliteration>The run and swim in (2) will be replaced by +.>
The neural network model coupled to the edit sequence in this embodiment includes an original encoderAnd decoder->The method comprises the steps of carrying out a first treatment on the surface of the Wherein, original text encoder->For coding translation originals, respectively>And machine translation text->Decoder->With original encoder->Takes as input the output hidden state of (2) to generate the machine translation error sequence in autoregressive manner>
Specifically, given a source language inputAnd machine translation output->Original encoder->For coding translation originals, respectively>And machine translation text->Converting it into a corresponding hidden state vector:
wherein->Is composed of m layers of neural network coding layers, wherein +.>The method comprises n layers of neural network coding layers; a plurality of sequence coding models such as Transformer, CNN, RNN can be used.
DecoderIn the state of translation original text hidden-> And the already generated machine translation error sequence +.>For input, the machine translation error sequence +.>
Wherein W is word vector matrix of decoder output layer,>the decoding layer is formed by stacking k decoding layers and one output layer, and each decoding layer comprises three sub-networks: the context represents a network, an original text attention network and a machine translation text mixed attention network, and the working principle of each sub-network is as follows:
(1) The context means that the network updates the current state according to the generated vector;
(2) The textual Attention network uses a Multi-Head Attention mechanism (Multi-Head Attention) to update the hidden state vector at the current time:wherein->Representing the state of translation origin in hidden->Representing a hidden state vector at the current moment;
(3) Machine translation of hidden state vector at current moment of mixed attention network updated with original attention networkAn organic translation text hidden state->For input, updated state vector is outputWherein->Is a mixed coefficient for measuring the importance degree of the machine translation on the decoding model, and the mixed coefficient can be used for relieving the translation error problem brought by a machine translation system, namely +.>The specific calculation mode of (2) is as follows: />Wherein->Representing a sigmoid activation function, W and b are model parameters, +.>Is the initial mixing coefficient, d represents the input state vector +.>Is a dimension of (c).
To sum up, given the instance inputs (S, MT, E), an optimized loss function for model training is constructed:
wherein (1)>For the initial character +.>For translating the length of the error sequence E, +.>For translating textual hidden state vector, +.>For machine translation hidden state vector, +.>The machine translation error sequence that has been generated.
Training the neural network model of the dual-to-edit sequence according to the constructed optimized loss function, training the neural network model in small batches, and training the neural network model by using an optimizer (such as Adam, SGD and the like) based on random gradient descent.
After model training, a neural network conversion model formula of the dual-to-edit sequence is obtained as follows:
wherein (1)>Representing a set of all training set instances, which employs fine-grained machine post-translational verification data (whereby training data may be constructed for model optimization based on corresponding error types),representing translation originals,/->Representing machine translation->Representing a machine translation error sequence,/->Is a model parameter.
After model training, the original text is translatedAnd machine translation text->As input, output of the machine translation error sequence corresponding to the machine translation text +.>(machine turning error sequence->Each term is represented by a quadruple:type), finally, according to the machine-turning error sequence +.>The error type, the position information and the text editing method based on the error type, and the automatic translation restoration is realized by correcting the machine translation through deleting and/or inserting text editing actions.
As shown in table two, the text editing method based on the error type is as follows: if the error type is mistranslation, excessive translation, grammar error, professional vocabulary error and fluency error, deleting the wrong translation and inserting the correct translation; if the error type is miss-translated, then the correct translation is inserted.
And (II) table: error type and corresponding editing operation
It should be noted that in the actual verification process (based on the real data), some error often needs to be modified multiple times, and in this case, the operations of "delete" and "insert" need to be called multiple times to correct the translation.
Example 2
As shown in fig. 3, the present embodiment provides a machine translation automatic post-verification system, which includes:
the model building module is used for training a neural network model of the dual-to-edit sequence by adopting fine-granularity machine translated verification data as training data to obtain a neural network conversion model of the dual-to-edit sequence; the fine-granularity machine translated verification data is obtained by manually marking translation errors of phrase levels on a machine translation result;
the machine-turning error prediction module is used for translating the original text based on a neural network conversion model of a dual-to-edit sequenceAnd machine translation text->As input, output of the machine translation error sequence corresponding to the machine translation text +.>
Automatic machine translation text restoration module based on machine turning error sequenceTranslation of text->Converting into target translation, wherein the machine-turned error sequence +.>Each term is represented by a quadruple: />Types.
It should be noted that the structure and/or principle of each module corresponds to the content of the automatic post-verification method for machine translation based on translation error correction described in embodiment 1, and thus will not be described herein.
It should be noted that, it should be understood that the division of each module of the above system is merely a division of a logic function, and may be fully or partially integrated into a physical entity in actual implementation, or may be physically separated, and the modules may be fully implemented in a form of software called by a processing element; or can be realized in hardware; the method can also be realized in a form of calling software by a processing element, and the method can be realized in a form of hardware by a part of modules. For example, a module may be a processing element that is set up separately, may be implemented in a chip of an apparatus, may be stored in a memory of the apparatus in the form of program codes, may be called by a processing element of the apparatus and perform functions of a module, and may be implemented similarly. In addition, all or part of the modules can be integrated together or can be independently implemented. The processing element described herein may be an integrated circuit having signal processing capabilities. In implementation, each step of the above method or each module above may be implemented by an integrated logic circuit of hardware in a processor element or an instruction in a software form.
For example, the modules above may be one or more integrated circuits configured to implement the methods above, such as: one or more specific integrated circuits, or one or more microprocessors, or one or more field programmable gate arrays, etc. For another example, when a module above is implemented in the form of a processing element scheduler code, the processing element may be a general purpose processor, such as a central processing unit or other processor that may invoke the program code. For another example, the modules may be integrated together and implemented in a system-on-chip form.
Example 3
The present embodiment provides a computer-readable storage medium having stored thereon a computer program that is executed by a processor to implement the machine translation automatic post-verification method based on translation error correction provided in embodiment 1. Those skilled in the art will appreciate that: all or part of the steps of implementing the method provided in embodiment 1 may be implemented by hardware associated with a computer program, where the computer program may be stored in a computer readable storage medium, and when executed, the program performs steps including the method provided in embodiment 1; and the storage medium includes: various media that can store program code, such as ROM, RAM, magnetic or optical disks.
Example 4
As shown in fig. 4, this embodiment provides a machine translation automatic post-verification device, including: a processor and a memory; the memory is used for storing a computer program; the processor is connected to the memory, and is configured to execute the computer program stored in the memory, so that the automatic post-machine-translation verification device performs the automatic post-machine-translation verification method based on translation error correction described in embodiment 1.
Specifically, the memory includes: various media capable of storing program codes, such as ROM, RAM, magnetic disk, U-disk, memory card, or optical disk.
Preferably, the processor may be a general-purpose processor, including a central processor, a network processor, etc.; but also digital signal processors, application specific integrated circuits, field programmable gate arrays or other programmable logic devices, discrete gate or transistor logic devices, discrete hardware components.
The foregoing is a preferred embodiment of the present invention. It should be noted that those skilled in the art may make several modifications without departing from the design principles and technical solutions of the present invention, and these modifications should also be considered as the protection scope of the present invention.

Claims (10)

1. A machine translation automatic post-verification method based on translation error correction, comprising:
(1) Training a neural network model of the dual-to-edit sequence by adopting fine-granularity machine translated verification data as training data to obtain a neural network conversion model of the dual-to-edit sequence;
(2) Based on the neural network conversion model of the dual-to-edit sequence obtained in the step (1) to translate the original textAnd machine translation text->As input, inputMachine translation error sequence corresponding to outgoing translation text +.>
(3) According to the sequence of machine turning errorsTranslation of text->Converting into a target translation;
the fine-granularity machine translated verification data is obtained by manually marking translation errors of phrase levels on a machine translation result.
2. The automatic post-verification method for machine translation based on translation error correction according to claim 1, wherein: the fine-grained machine post-translation verification data is synthesized by manual labelingMachine translation error sequence obtained from three sequences +.>,/>Wherein->Represents the nth machine turning error, +.>Representing the length of the machine-turning error sequence;
which represents the translation context; />Which represents the machine translation;which represents a post-verification translation, wherein +.>Representing translation origin, machine translation and post-verification translation respectively +.>Personal word (s)/(s)>Respectively represent->The length of the three sequences.
3. The automatic post-verification method for machine translation based on translation error correction according to claim 2, wherein: the machine translation error sequenceIs represented by a quadruple: mechanically turning the phrase sequence, the modified sequence, the original text corresponding sequence and the error type; the error types include mistranslations, missed translations, transitional translations, professional vocabulary errors, grammar errors, and/or fluency errors.
4. A machine translation automatic post-verification method based on translation error correction according to claim 3, wherein: the neural network model of the dual-to-edit sequence in the step (1) comprises an original text encoderMachine translation->And decoder->The method comprises the steps of carrying out a first treatment on the surface of the Wherein, original text encoder->Machine translation->For coding translation originals, respectively>And machine translation text->Decoder->With original encoder->Machine translation->Takes as input the output hidden state of (2) to generate the machine translation error sequence in autoregressive manner>
5. The automatic post-translational error correction based machine translation verification method of claim 4, wherein said textual encoderMachine translation->For coding translation originals, respectively>And machine translation text->Converting it into a corresponding hidden state vector:
wherein->Is composed of m layers of neural network coding layers, wherein +.>Is composed of n nerve network coding layers.
6. A machine translation automatic post-verification method based on translation error correction according to claim 5, wherein said decoderIn the state of translation original text hidden->Hidden state of machine translation>And the already generated machine translation error sequence +.>For input, the machine translation error sequence +.> Wherein W is word vector matrix of decoder output layer,>is formed by stacking k decoding layers and one output layer.
7. The automatic post-translation verification based on translation error correction method according to claim 6, wherein said decoding layer comprises three sub-networks: context means network, textual attention network and machine translation textual mixed attention network;
the context represents that the network updates the current state according to the generated vector;
the textual attention network uses a multi-headed attention mechanism to update the hidden state vector at the current time:wherein->Show the state of translation original text hidden->Representing a hidden state vector at the current moment;
the machine translation text mixed attention network updates the hidden state vector of the current moment with the original text attention networkAn organic translation text hidden state->For input, updated state vector is outputWherein->Is a mixed coefficient for measuring the importance of a machine translation to a decoding model, said +.>The calculation mode of (2) is as follows:
wherein->Representing a sigmoid activation function, W and b are model parameters, +.>Initial mixing coefficient, d represents the input state vector +.>Is a dimension of (c).
8. The automatic post-verification of machine translation based on translation error correction of claim 7, wherein (1) further comprises a given instance input (S, MT, E), constructing a model trained optimization loss function:wherein (1)>For the initial character +.>For translating the length of the error sequence E, +.>For translating textual hidden state vector, +.>For machine translation hidden state vector, +.>Is an already generated machine translation error sequence.
9. The automatic post-verification method for machine translation based on translation error correction according to claim 8, wherein the neural network conversion model formula of the dual-to-edit sequence in (1) is expressed as follows:wherein (1)>Representing the set of all training set instances, which uses the fine-grained machine post-translational verification data, +.>Representing translation originals,/->Representing machine translation->Representing a machine translation error sequence,/->Is a model parameter.
10. The automatic post-verification method for machine translation based on translation error correction according to claim 9, wherein the specific method of (3) is as follows: according to the sequence of machine turning errorsThe error type, the position information and the text editing method based on the error type, and the automatic translation restoration is realized by correcting the machine translation through deleting and/or inserting text editing actions.
CN202311559241.5A 2023-11-22 2023-11-22 Automatic machine translation post-verification method based on translation error correction Active CN117273027B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202311559241.5A CN117273027B (en) 2023-11-22 2023-11-22 Automatic machine translation post-verification method based on translation error correction

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202311559241.5A CN117273027B (en) 2023-11-22 2023-11-22 Automatic machine translation post-verification method based on translation error correction

Publications (2)

Publication Number Publication Date
CN117273027A true CN117273027A (en) 2023-12-22
CN117273027B CN117273027B (en) 2024-04-30

Family

ID=89220006

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202311559241.5A Active CN117273027B (en) 2023-11-22 2023-11-22 Automatic machine translation post-verification method based on translation error correction

Country Status (1)

Country Link
CN (1) CN117273027B (en)

Citations (11)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20190129946A1 (en) * 2017-10-30 2019-05-02 Sdl Inc. Fragment Recall and Adaptive Automated Translation
CN110069790A (en) * 2019-05-10 2019-07-30 东北大学 It is a kind of by translation retroversion to machine translation system and method literally
CN111597778A (en) * 2020-04-15 2020-08-28 哈尔滨工业大学 Method and system for automatically optimizing machine translation based on self-supervision
CN112287696A (en) * 2020-10-29 2021-01-29 语联网(武汉)信息技术有限公司 Post-translation editing method and device, electronic equipment and storage medium
CN112347795A (en) * 2020-10-04 2021-02-09 北京交通大学 Machine translation quality evaluation method, device, equipment and medium
CN113408307A (en) * 2021-07-14 2021-09-17 北京理工大学 Neural machine translation method based on translation template
CN115345181A (en) * 2022-07-04 2022-11-15 中国科学院自动化研究所 Training method of neural machine translation model, translation method and device
CN116502653A (en) * 2022-01-18 2023-07-28 华为技术有限公司 Translation quality evaluation method, electronic device, chip and readable storage medium
CN116822534A (en) * 2023-07-06 2023-09-29 四川语言桥信息技术有限公司 Fine granularity characteristic-based machine turning evaluation index interpretation method, interpreter model and computer readable storage medium
CN116956946A (en) * 2023-07-14 2023-10-27 上海一者信息科技有限公司 Machine translation text fine granularity error type identification and positioning method
CN116992892A (en) * 2023-07-06 2023-11-03 四川语言桥信息技术有限公司 Method, system and readable storage medium for improving APE model based on data enhancement and multitasking training

Patent Citations (11)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20190129946A1 (en) * 2017-10-30 2019-05-02 Sdl Inc. Fragment Recall and Adaptive Automated Translation
CN110069790A (en) * 2019-05-10 2019-07-30 东北大学 It is a kind of by translation retroversion to machine translation system and method literally
CN111597778A (en) * 2020-04-15 2020-08-28 哈尔滨工业大学 Method and system for automatically optimizing machine translation based on self-supervision
CN112347795A (en) * 2020-10-04 2021-02-09 北京交通大学 Machine translation quality evaluation method, device, equipment and medium
CN112287696A (en) * 2020-10-29 2021-01-29 语联网(武汉)信息技术有限公司 Post-translation editing method and device, electronic equipment and storage medium
CN113408307A (en) * 2021-07-14 2021-09-17 北京理工大学 Neural machine translation method based on translation template
CN116502653A (en) * 2022-01-18 2023-07-28 华为技术有限公司 Translation quality evaluation method, electronic device, chip and readable storage medium
CN115345181A (en) * 2022-07-04 2022-11-15 中国科学院自动化研究所 Training method of neural machine translation model, translation method and device
CN116822534A (en) * 2023-07-06 2023-09-29 四川语言桥信息技术有限公司 Fine granularity characteristic-based machine turning evaluation index interpretation method, interpreter model and computer readable storage medium
CN116992892A (en) * 2023-07-06 2023-11-03 四川语言桥信息技术有限公司 Method, system and readable storage medium for improving APE model based on data enhancement and multitasking training
CN116956946A (en) * 2023-07-14 2023-10-27 上海一者信息科技有限公司 Machine translation text fine granularity error type identification and positioning method

Non-Patent Citations (2)

* Cited by examiner, † Cited by third party
Title
YAFU LI 等: "On compositional generalization of neural machine translation", 《COMPUTATION AND LANGUAGE》, pages 1 - 14 *
周纯: "神经网络机器翻译技术背景下新闻文本译后编辑研究", 《品位经典》, no. 4, pages 29 - 30 *

Also Published As

Publication number Publication date
CN117273027B (en) 2024-04-30

Similar Documents

Publication Publication Date Title
CN109446534B (en) Machine translation method and device
CN110069790B (en) Machine translation system and method for contrasting original text through translated text retranslation
CN108170686B (en) Text translation method and device
CN107273356B (en) Artificial intelligence based word segmentation method, device, server and storage medium
CN113811946A (en) End-to-end automatic speech recognition of digital sequences
CN110532573A (en) A kind of interpretation method and system
CN108132932B (en) Neural machine translation method with replication mechanism
CN109522403A (en) A kind of summary texts generation method based on fusion coding
WO2021127817A1 (en) Speech synthesis method, device, and apparatus for multilingual text, and storage medium
WO2021082427A1 (en) Rhythm-controlled poem generation method and apparatus, and device and storage medium
CN111177348B (en) Training method and device for problem generation model, electronic equipment and storage medium
WO2022088570A1 (en) Method and apparatus for post-editing of translation, electronic device, and storage medium
CN111144140B (en) Zhongtai bilingual corpus generation method and device based on zero-order learning
CN110807335A (en) Translation method, device, equipment and storage medium based on machine learning
CN110222350A (en) By bilingual predefined translation to the method for incorporating neural Machine Translation Model
CN114925170B (en) Text proofreading model training method and device and computing equipment
US11615247B1 (en) Labeling method and apparatus for named entity recognition of legal instrument
CN115906815A (en) Error correction method and device for modifying one or more types of wrong sentences
CN116822464A (en) Text error correction method, system, equipment and storage medium
CN113626563A (en) Method and electronic equipment for training natural language processing model and natural language processing
CN117195922A (en) Human-in-loop neural machine translation method, system and readable storage medium
Zhao et al. Tibetan multi-dialect speech recognition using latent regression Bayesian network and end-to-end mode
CN117273027B (en) Automatic machine translation post-verification method based on translation error correction
CN114330375A (en) Term translation method and system based on fixed paradigm
CN115374784A (en) Chinese named entity recognition method based on multi-mode information selective fusion

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant