CN116822534A - Fine granularity characteristic-based machine turning evaluation index interpretation method, interpreter model and computer readable storage medium - Google Patents

Fine granularity characteristic-based machine turning evaluation index interpretation method, interpreter model and computer readable storage medium Download PDF

Info

Publication number
CN116822534A
CN116822534A CN202310819638.7A CN202310819638A CN116822534A CN 116822534 A CN116822534 A CN 116822534A CN 202310819638 A CN202310819638 A CN 202310819638A CN 116822534 A CN116822534 A CN 116822534A
Authority
CN
China
Prior art keywords
score
sequence
attention
machine
text
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN202310819638.7A
Other languages
Chinese (zh)
Inventor
朱宪超
胡刚
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Sichuan Lan Bridge Information Technology Co ltd
Original Assignee
Sichuan Lan Bridge Information Technology Co ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Sichuan Lan Bridge Information Technology Co ltd filed Critical Sichuan Lan Bridge Information Technology Co ltd
Priority to CN202310819638.7A priority Critical patent/CN116822534A/en
Publication of CN116822534A publication Critical patent/CN116822534A/en
Pending legal-status Critical Current

Links

Landscapes

  • Machine Translation (AREA)

Abstract

The invention belongs to the technical field of machine translation, and provides an interpretation method, an interpreter model and a computer readable storage medium of a machine turning evaluation index based on fine granularity characteristics. According to the method, the problem of insufficient interpretability of the machine turning quality assessment model is solved through the interpretation method/the interpreter model of the machine turning assessment index with the fine granularity characteristics, the fine granularity interpretability characteristics suitable for the machine turning scene are provided for the quality assessment score, and the characteristics are expected to be applied to the artificial PE stage of the machine turning and improve the efficiency of the artificial PE.

Description

Fine granularity characteristic-based machine turning evaluation index interpretation method, interpreter model and computer readable storage medium
Technical Field
The invention belongs to the technical field of machine translation, and particularly relates to an interpretation method, an interpreter model and a computer readable storage medium of a machine turning evaluation index based on fine granularity characteristics.
Background
In the traditional manual evaluation means, an evaluator usually evaluates the quality of the machine translation from three aspects of translation sufficiency, translation fluency and translation readability, and in order to obtain more rapid evaluation, some researchers propose rule-based evaluation indexes such as BLEU. The BLEU index is evaluated based on co-occurrence rate of words or N-gram, and contains a sufficiency index in manual evaluation. In recent years, in order to better evaluate the quality of the translation of the machine, another part of researchers put forward evaluation indexes such as BertScore, moverScore and COMET based on a pre-training model, and the indexes can give consideration to the sufficiency and fluency of the translation in evaluating the quality of the translation of the machine, so that the evaluation indexes are more advanced evaluation indexes.
However, at the present stage, more advanced machine translation evaluation indexes such as BertScore are not used in the evaluation of machine translations on a large scale, because: these evaluation indexes based on the pre-training model lack interpretability, but are of a black box structure, and are difficult to form reasonable interpretation.
To solve this problem, researchers have attempted to interpret the results of the machine-turning evaluation model using an interpreter model such as LIME, SHAP, etc. in machine learning. However, these methods generally only give an explanation from the word position level, and do not give information more conforming to other explanation types of the machine-turning scene, such as types of errors; and an interpreter model based on a simple linear model cannot express fine granularity characteristics in a machine translation text quality evaluation scene.
Disclosure of Invention
The invention aims to provide an interpretation method of a turning evaluation index based on fine granularity characteristics, which aims to solve the technical problems existing in the prior art.
In order to achieve the above purpose, the technical scheme adopted by the invention is as follows:
an interpretation method of machine turning evaluation indexes based on fine granularity characteristics comprises the following steps:
step S1: turning original text sequence S= (S) 1 ,s 2 ,...,s n ) Machine translated text sequence h= (H) 1 ,h 2 ,...,h m ) And a quality assessment score Q, spliced together using labels;
step S2: inputting the data obtained in the step S1 into a pre-training model, and calculating to obtain a coding sequence:
T=Encoder(S;H;Q)
wherein, the Encoder is a pre-training model, and T is a context label obtained by calculation through the pre-training model;
step S3: dividing the coding sequence according to the input position to obtain context expression T of the original text and the translated text text And the context of the quality assessment score represents T score
Step S4: calculating contextual representations T of originals and translations using an attention mechanism text And the context of the quality assessment score represents T score Is a matrix of attention weights:
attention=softrmax(T text ⊙T score );
step S5: carrying out mean value pooling on the attention weight matrix to obtain a one-dimensional relative importance score State:
State=MeanPooling(attention).
step S6: marking the start and stop positions of each position related to the quality score by using a sequence marked tag, and marking the error type for each position to obtain a fine-granularity interpretable feature tag;
step S7: based on the relative importance score State and the fine granularity interpretable feature tag, using CRF calculation to obtain a sequence tag y= (Y) 1 ,y 2 ,...,y n+m ):
Y=CRF(State)
In one embodiment, SEP tag splicing is used in step S1.
In one embodiment, the attention mechanism in step S4 employs one of a point-by-point attention mechanism, a connection attention mechanism, and a bilinear attention mechanism.
In one embodiment, a BIO tag is used in step S6.
In one embodiment, the fine-grained interpretable feature tag of step S6 includes:
b-miss translation starting position and I-miss translation intermediate position;
B-misT transliteration start position, I-misT transliteration intermediate position;
b-over-translation start position, I-over-translation intermediate position.
In one embodiment, the machine-turned original sequence s= (S) in step S1 1 ,s 2 ,...,s n ) Machine translated text sequence h= (H) 1 ,h 2 ,...,h m ) And the quality evaluation score Q is obtained through machine turning quality evaluation model reasoning.
In one embodiment, the error type mining method is as follows: first, the sequence s= (S) of the machine original text is inferred by the attention mechanism 1 ,s 2 ,…,s n ) Mechanism translation sequence h= (H) 1 ,h 2 ,...,h m ) And an error segment of the quality assessment score Q, and is represented as a signal of high or low weight in the attention matrix; then, the signals with high and low weight values are mapped into transition state signals of 0-n, wherein n represents the number of error types.
In order to achieve the above object, the present invention further provides an interpreter model for a machine turning evaluation index based on fine granularity characteristics, including:
and (3) splicing modules: turning original text sequence S= (S) 1 ,s 2 ,...,s n ) Machine translated text sequence h= (H) 1 ,h 2 ,...,h m ) And a quality assessment score Q, spliced together using labels;
coding sequence module: inputting the obtained data into a pre-training model, and calculating to obtain a coding sequence:
T=Encoder(S;H;Q)
wherein, the Encoder is a pre-training model, and T is a context label obtained by calculation through the pre-training model;
the context representation module: dividing the coding sequence according to the input position to obtain context expression T of the original text and the translated text text And the context of the quality assessment score represents T score
Note weight module: using attention machinesProducing, calculating context representation T of original text and translated text text And the context of the quality assessment score represents T score Is a matrix of attention weights:
attention=softmax(T text ⊙T score );
a relative importance score module: carrying out mean value pooling on the attention weight matrix to obtain a one-dimensional relative importance score State:
State=MeanPooling(attention);
fine granularity interpretability feature tag module: marking the start and stop positions of each position related to the quality score by using a sequence marked tag, and marking the error type for each position to obtain a fine-granularity interpretable feature tag;
sequence tag module: based on the relative importance score State and the fine granularity interpretable feature tag, using CRF calculation to obtain a sequence tag y= (Y) 1 ,y 2 ,...,y n+m ):
Y=CRF(State)。
In order to achieve the above object, the present invention also provides a computer-readable storage medium having stored thereon a computer program that is executed by a processor to implement the explained method of the machine turning evaluation index based on fine granularity characteristics.
Compared with the prior art, the invention has the following beneficial effects:
according to the method, the problem of insufficient interpretability of the machine turning quality assessment model is solved through the interpretation method/the interpreter model of the machine turning assessment index with the fine granularity characteristics, the fine granularity interpretability characteristics suitable for the machine turning scene are provided for the quality assessment score, and the characteristics are expected to be applied to the artificial PE stage of the machine turning and improve the efficiency of the artificial PE.
Drawings
FIG. 1 is a schematic flow chart of the present invention-example 1.
Fig. 2 is a schematic diagram of a model structure of an interpreter model in embodiment 2 of the present invention.
Fig. 3 is a schematic block diagram of an interpreter model in the present invention-embodiment 2.
Detailed Description
The present invention will be further described in detail with reference to examples so as to enable those skilled in the art to more clearly understand and understand the present invention. It should be understood that the following specific embodiments are only for explaining the present invention, and it is convenient to understand that the technical solutions provided by the present invention are not limited to the technical solutions provided by the following embodiments, and the technical solutions provided by the embodiments should not limit the protection scope of the present invention.
It should be noted that the illustrations provided in the following embodiments merely illustrate the basic concept of the present invention by way of illustration, so that only the components related to the present invention are shown in the drawings and are not drawn according to the number, shape and size of the components in actual implementation, the form, number and proportion of each component in actual implementation may be arbitrarily changed, and the layout of the components may be more complicated.
Example 1
As shown in fig. 1, the embodiment provides an interpretation method of a turning evaluation index based on fine-grained characteristics, which implements interpretation of the turning evaluation index based on fine-grained characteristics, and specifically includes the following implementation steps:
step S1: turning original text sequence S= (S) 1 ,s 2 ,...,s n ) Machine translated text sequence h= (H) 1 ,h 2 ,...,h m ) And a quality assessment score Q, stitched together using SEP tags, wherein the machine-inverted sequence s= (S) 1 ,s 2 ,...,s n ) Machine translated text sequence h= (H) 1 ,h 2 ,...,h m ) And the quality evaluation score Q is obtained through machine turning quality evaluation model reasoning. Examples: the original data are original sentences: ilike NBA, translated sentence: i like basketball, HTER score: 0.32, the preprocessing goes to step S1 (including word segmentation and data concatenation operations): [ CLS ]]I like NBA,[SEP]I like basketball, [ SEP ]]0.32。
Step S2: inputting the data obtained in the step S1 into a pre-training model, and calculating to obtain a coding sequence:
T=Encoder(S;H;Q)
wherein the Encoder is a pre-trained model, such as: XLM-R, other training models may also be selected by those skilled in the art, and are not particularly limited herein; t is a context label obtained through calculation of a pre-training model; further, the method comprises the steps of deducing error fragments of the original text sequence, the machine translation text sequence and the quality assessment score through a pre-training model, and exploring error types through the error fragments.
Step S3: dividing the code sequence according to the input position to obtain the context expression T of the original text and the translated text text And the context of the quality assessment score represents T score
Step S4: calculating a contextual representation T of the original and translated text using one of a point-multiply attention mechanism, a join attention mechanism, and a bilinear attention mechanism text And the context of the quality assessment score represents T score Is a matrix of attention weights:
attention=softmax(T text ⊙T score );
in combination with the above example, hidden layer representations are generated by pre-training the model, and attention weights are calculated through the attention layers, the model having been able to focus on the corresponding anomalies between "NBA" and "basketball";
step S5: carrying out mean value pooling on the attention weight matrix to obtain a one-dimensional relative importance score State:
State=MeanPooling(attention).
step S6: marking the start and stop positions of each position related to the quality score by using a BIO label marked by a sequence, and marking the error type for each position to obtain a fine-granularity interpretable feature label, wherein the label comprises three pairs, and the method comprises the following steps of: (1) B-miss translation starting position and I-miss translation intermediate position; (2) B-misT transliteration start position, I-misT transliteration intermediate position; (3) B-over translation starting position, I-over translation intermediate position; the error type mining method comprises the following steps: in steps S4 and S5, deducing error fragments related to the machine-turned text sequence, the machine-translated text sequence and the quality evaluation score through an attention mechanism, and representing the error fragments as signals with high and low weight values in an attention matrix, and mapping the signals with high and low weight values into transition state signals of 0-n through a subsequent neural network model in steps S6 and S7, wherein n represents the number of error types, so that the error types can be extracted from the error fragments.
Step S7: based on the relative importance score State and the fine granularity interpretable feature tag, a sequence tag y= (Y) is calculated using CRF (Conditional Random Field ) 1 ,y 2 ,...,y n+m ):
Y=CRF(State);
In combination with the above example, the corresponding error location tag is obtained through the processing of the above steps: 0 0 0B-misT 0 0.
In the prior art, the interpretation evaluation result does not calculate (explore) the error type of the error fragment, and the technical scheme provided by the invention continues to calculate the error type corresponding to the error fragment while marking the error fragment.
Example 2
As shown in fig. 2 and 3, the present embodiment provides an interpreter model of a machine turning evaluation index based on fine granularity characteristics, the interpreter model including:
and (3) splicing modules: turning original text sequence S= (S) 1 ,s 2 ,...,s n ) Machine translated text sequence h= (H) 1 ,h 2 ,...,h m ) And a quality assessment score Q, stitched together using SEP tags; wherein, the machine-turned original text sequence S= (S) 1 ,s 2 ,...,s n ) Machine translated text sequence h= (H) 1 ,h 2 ,...,h m ) The quality evaluation score Q is obtained through reasoning of a machine turning quality evaluation model;
coding sequence module: inputting the obtained data into a pre-training model, and calculating to obtain a coding sequence:
T=Encoder(S;H;Q)
wherein, the Encoder is a pre-training model, and T is a context label obtained by calculation through the pre-training model;
the context representation module: dividing the code sequence according to the input position to obtain the context expression T of the original text and the translated text text And the context of the quality assessment score represents T score
Note weight module: calculating the contextual representation T of the original and translated text using a point-multiply attention mechanism, a join attention mechanism, or a bilinear attention mechanism text And the context of the quality assessment score represents T score Is a matrix of attention weights:
attention=softmax(T text ⊙T score );
a relative importance score module: carrying out mean value pooling on the attention weight matrix to obtain a one-dimensional relative importance score State:
State=MeanPooling(attention).
fine granularity interpretability feature tag module: marking the start and stop positions of each position related to the quality score by using a BIO label marked by a sequence, and marking the error type for each position to obtain a fine-granularity interpretable feature label: b-miss translation starting position and I-miss translation intermediate position; B-misT transliteration start position, I-misT transliteration intermediate position; b-over translation starting position, I-over translation intermediate position;
sequence tag module: based on the relative importance score State and the fine granularity interpretable feature tag, a sequence tag y= (Y) is calculated using CRF 1 ,y 2 ,...,y n+m ):
Y=CRF(State)。
It should be noted that the structure and/or principle of each module corresponds to the content of the explanation method of the turning evaluation index based on the fine granularity feature described in embodiment 1, and thus will not be described herein.
It should be noted that, it should be understood that the division of each module of the above system is merely a division of a logic function, and may be fully or partially integrated into a physical entity in actual implementation, or may be physically separated, and the modules may be fully implemented in a form of software called by a processing element; or can be realized in hardware; the method can also be realized in a form of calling software by a processing element, and the method can be realized in a form of hardware by a part of modules. For example, a module may be a processing element that is set up separately, may be implemented in a chip of an apparatus, may be stored in a memory of the apparatus in the form of program codes, may be called by a processing element of the apparatus and perform functions of a module, and may be implemented similarly. In addition, all or part of the modules can be integrated together or can be independently implemented. The processing element described herein may be an integrated circuit having signal processing capabilities. In implementation, each step of the above method or each module above may be implemented by an integrated logic circuit of hardware in a processor element or an instruction in a software form.
For example, the modules above may be one or more integrated circuits configured to implement the methods above, such as: one or more specific integrated circuits, or one or more microprocessors, or one or more field programmable gate arrays, etc. For another example, when a module above is implemented in the form of a processing element scheduler code, the processing element may be a general purpose processor, such as a central processing unit or other processor that may invoke the program code. For another example, the modules may be integrated together and implemented in a system-on-chip form.
Example 3
The present embodiment provides a computer-readable storage medium having stored thereon a computer program that is executed by a processor to implement the interpretation method of the turning evaluation index based on the fine-grained feature provided in embodiment 1. Those skilled in the art will appreciate that: all or part of the steps of implementing the method provided in embodiment 1 may be implemented by hardware associated with a computer program, where the computer program may be stored in a computer readable storage medium, and when executed, the program performs steps including the method provided in embodiment 1; and the storage medium includes: various media that can store program code, such as ROM, RAM, magnetic or optical disks.
The foregoing is a preferred embodiment of the present invention. It should be noted that those skilled in the art may make several modifications without departing from the design principles and technical solutions of the present invention, and these modifications should also be considered as the protection scope of the present invention.

Claims (9)

1. The interpretation method of the turning evaluation index based on the fine granularity characteristics is characterized by comprising the following steps of:
step S1: turning original text sequence S= (S) 1 ,s 2 ,...,s n ) Machine translated text sequence h= (H) 1 ,h 2 ,...,h m ) And a quality assessment score Q, spliced together using labels;
step S2: inputting the data obtained in the step S1 into a pre-training model, and calculating to obtain a coding sequence:
T=Encoder(S;H;Q)
wherein, the Encoder is a pre-training model, and T is a context label obtained by calculation through the pre-training model;
step S3: dividing the coding sequence according to the input position to obtain context expression T of the original text and the translated text text And the context of the quality assessment score represents T score
Step S4: calculating contextual representations T of originals and translations using an attention mechanism text And the context of the quality assessment score represents T score Is a matrix of attention weights:
attention=softmax(T text ⊙T score );
step S5: carrying out mean value pooling on the attention weight matrix to obtain a one-dimensional relative importance score State:
State=MeanPooling(attention);
step S6: marking the start and stop positions of each position related to the quality score by using a sequence marked tag, and marking the error type for each position to obtain a fine-granularity interpretable feature tag;
step S7: based onThe relative importance score State and the fine granularity interpretable feature tag are calculated using CRF to obtain a sequence tag Y= (Y) 1 ,y 2 ,...,y n+m ):
Y=CRF(State)。
2. The interpretation method of the turning evaluation index based on the fine granularity characteristics according to claim 1, wherein the interpretation method comprises the following steps: in the step S1, SEP label stitching is used.
3. The method according to claim 2, wherein the attention mechanism in step S4 is one of a point-by-point attention mechanism, a connection attention mechanism, and a bilinear attention mechanism.
4. The method for interpreting a turning evaluation index based on fine-grained characteristics according to claim 3, wherein a BIO tag is used in said step S6.
5. The method for interpreting a machine-turning evaluation index based on fine-grained feature according to claim 4, wherein the fine-grained interpretable feature tag in step S6 comprises:
b-miss translation starting position and I-miss translation intermediate position;
B-misT transliteration start position, I-misT transliteration intermediate position;
b-over-translation start position, I-over-translation intermediate position.
6. The method for interpreting a machine-turning evaluation index based on fine granularity features as claimed in claim 5, wherein the machine-turning original text sequence s= (S) in step S1 1 ,s 2 ,...,s n ) Machine translated text sequence h= (H) 1 ,h 2 ,...,h m ) And the quality evaluation score Q is obtained through machine turning quality evaluation model reasoning.
7. The interpretation method of the turning evaluation index based on the fine granularity feature according to claim 6, characterized in that the error type mining method is as follows; first, the machine-turned original sequence s= (S) is inferred by the attention mechanism 1 ,s 2 ,...,s n ) Machine translated text sequence h= (H) 1 ,h 2 ,...,h m ) And an error segment of the quality assessment score Q, and is represented as a signal of high or low weight in the attention matrix; then, the signals with high and low weight values are mapped into transition state signals of 0-n, wherein n represents the number of error types.
8. An interpreter model of a machine turning evaluation index based on fine granularity characteristics, which is characterized by comprising:
and (3) splicing modules: turning original text sequence S= (S) 1 ,s 2 ,...,s n ) Machine translated text sequence h= (H) 1 ,h 2 ,...,h m ) And a quality assessment score Q, spliced together using labels;
coding sequence module: inputting the obtained data into a pre-training model, and calculating to obtain a coding sequence:
T=Encoder(S;H;Q)
wherein, the Encoder is a pre-training model, and T is a context label obtained by calculation through the pre-training model;
the context representation module: dividing the coding sequence according to the input position to obtain context expression T of the original text and the translated text text And the context of the quality assessment score represents T score
Note weight module: calculating contextual representations T of originals and translations using an attention mechanism text And the context of the quality assessment score represents T score Is a matrix of attention weights:
attention=softrmax(T text ⊙T score );
a relative importance score module: carrying out mean value pooling on the attention weight matrix to obtain a one-dimensional relative importance score State:
State=MeanPooling(attention);
fine granularity interpretability feature tag module: marking the start and stop positions of each position related to the quality score by using a sequence marked tag, and marking the error type for each position to obtain a fine-granularity interpretable feature tag;
sequence tag module: based on the relative importance score State and the fine granularity interpretable feature tag, using CRF calculation to obtain a sequence tag y= (Y) 1 ,y 2 ,...,y n+m ):
Y=CRF(State)。
9. A computer-readable storage medium, on which a computer program is stored, characterized in that the computer program is executed by a processor to implement the interpretation method of the machine turning evaluation index based on fine-grained feature as set forth in any one of claims 1 to 7.
CN202310819638.7A 2023-07-06 2023-07-06 Fine granularity characteristic-based machine turning evaluation index interpretation method, interpreter model and computer readable storage medium Pending CN116822534A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202310819638.7A CN116822534A (en) 2023-07-06 2023-07-06 Fine granularity characteristic-based machine turning evaluation index interpretation method, interpreter model and computer readable storage medium

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202310819638.7A CN116822534A (en) 2023-07-06 2023-07-06 Fine granularity characteristic-based machine turning evaluation index interpretation method, interpreter model and computer readable storage medium

Publications (1)

Publication Number Publication Date
CN116822534A true CN116822534A (en) 2023-09-29

Family

ID=88118273

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202310819638.7A Pending CN116822534A (en) 2023-07-06 2023-07-06 Fine granularity characteristic-based machine turning evaluation index interpretation method, interpreter model and computer readable storage medium

Country Status (1)

Country Link
CN (1) CN116822534A (en)

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN117273027A (en) * 2023-11-22 2023-12-22 四川语言桥信息技术有限公司 Automatic machine translation post-verification method based on translation error correction
CN117436460A (en) * 2023-12-22 2024-01-23 武汉大学 Translation quality assessment method, device, equipment and storage medium

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN112347795A (en) * 2020-10-04 2021-02-09 北京交通大学 Machine translation quality evaluation method, device, equipment and medium
US20210174033A1 (en) * 2019-12-05 2021-06-10 Baidu Online Network Technology (Beijing) Co., Ltd. Method and apparatus for evaluating translation quality
CN114091480A (en) * 2021-11-18 2022-02-25 北京理工大学 Machine translation quality estimation method integrating and utilizing multiple pre-training models
CN114580437A (en) * 2022-03-02 2022-06-03 阿里巴巴(中国)有限公司 Training method of translation evaluation model, translation evaluation method, device and medium

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20210174033A1 (en) * 2019-12-05 2021-06-10 Baidu Online Network Technology (Beijing) Co., Ltd. Method and apparatus for evaluating translation quality
CN112347795A (en) * 2020-10-04 2021-02-09 北京交通大学 Machine translation quality evaluation method, device, equipment and medium
CN114091480A (en) * 2021-11-18 2022-02-25 北京理工大学 Machine translation quality estimation method integrating and utilizing multiple pre-training models
CN114580437A (en) * 2022-03-02 2022-06-03 阿里巴巴(中国)有限公司 Training method of translation evaluation model, translation evaluation method, device and medium

Non-Patent Citations (2)

* Cited by examiner, † Cited by third party
Title
BIN NI 等: "SynXLM-R: Syntax-Enhanced XLM-R in Translation Quality Estimation", NATURAL LANGUAGE PROCESSING AND CHINESE COMPUTING, pages 27 - 40 *
孟庆晔: "基于预训练语言模型的翻译质量估计研究", 中国优秀硕士学位论文全文数据库, pages 138 - 3146 *

Cited By (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN117273027A (en) * 2023-11-22 2023-12-22 四川语言桥信息技术有限公司 Automatic machine translation post-verification method based on translation error correction
CN117273027B (en) * 2023-11-22 2024-04-30 四川语言桥信息技术有限公司 Automatic machine translation post-verification method based on translation error correction
CN117436460A (en) * 2023-12-22 2024-01-23 武汉大学 Translation quality assessment method, device, equipment and storage medium
CN117436460B (en) * 2023-12-22 2024-03-12 武汉大学 Translation quality assessment method, device, equipment and storage medium

Similar Documents

Publication Publication Date Title
CN116822534A (en) Fine granularity characteristic-based machine turning evaluation index interpretation method, interpreter model and computer readable storage medium
CN110111864B (en) Medical report generation system and method based on relational model
CN110427623A (en) Semi-structured document Knowledge Extraction Method, device, electronic equipment and storage medium
CN107832400A (en) A kind of method that location-based LSTM and CNN conjunctive models carry out relation classification
CN107220506A (en) Breast cancer risk assessment analysis system based on deep convolutional neural network
CN113255320A (en) Entity relation extraction method and device based on syntax tree and graph attention machine mechanism
CN114547298B (en) Biomedical relation extraction method, device and medium based on combination of multi-head attention and graph convolution network and R-Drop mechanism
CN114118065A (en) Chinese text error correction method and device in electric power field, storage medium and computing equipment
CN115062140A (en) Method for generating abstract of BERT SUM and PGN fused supply chain ecological district length document
WO2017132545A1 (en) Systems and methods for generative learning
CN112069827B (en) Data-to-text generation method based on fine-grained subject modeling
CN115545041B (en) Model construction method and system for enhancing semantic vector representation of medical statement
CN114528374A (en) Movie comment emotion classification method and device based on graph neural network
CN113920379A (en) Zero sample image classification method based on knowledge assistance
CN114281982B (en) Book propaganda abstract generation method and system adopting multi-mode fusion technology
Li et al. Auto completion of user interface layout design using transformer-based tree decoders
CN112836482A (en) Method and device for generating problems by sequence generation model based on template
WO2023134676A1 (en) Method and system for automatically formulating optimization problem using machine learning
CN117521074A (en) Knowledge-driven-based pre-training model software vulnerability entity recognition device and method
CN115659981A (en) Named entity recognition method based on neural network model
CN115270792A (en) Medical entity identification method and device
CN114372467A (en) Named entity extraction method and device, electronic equipment and storage medium
CN114510569A (en) Chemical emergency news classification method based on Chinesebert model and attention mechanism
CN113486180A (en) Remote supervision relation extraction method and system based on relation hierarchy interaction
Xie et al. Enhancing multimodal deep representation learning by fixed model reuse

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination