WO2022057116A1 - 一种基于Transformer深度学习模型的多语种地名词根汉译方法 - Google Patents
一种基于Transformer深度学习模型的多语种地名词根汉译方法 Download PDFInfo
- Publication number
- WO2022057116A1 WO2022057116A1 PCT/CN2020/136009 CN2020136009W WO2022057116A1 WO 2022057116 A1 WO2022057116 A1 WO 2022057116A1 CN 2020136009 W CN2020136009 W CN 2020136009W WO 2022057116 A1 WO2022057116 A1 WO 2022057116A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- place name
- root
- chinese
- language
- place
- Prior art date
Links
- 238000000034 method Methods 0.000 title claims abstract description 33
- 238000013136 deep learning model Methods 0.000 title claims description 13
- 238000013519 translation Methods 0.000 claims abstract description 47
- 239000013598 vector Substances 0.000 claims abstract description 33
- 238000012549 training Methods 0.000 claims abstract description 30
- 238000000605 extraction Methods 0.000 claims abstract description 25
- 230000007246 mechanism Effects 0.000 claims description 30
- 238000013528 artificial neural network Methods 0.000 claims description 26
- 238000012545 processing Methods 0.000 claims description 13
- 238000012360 testing method Methods 0.000 claims description 7
- 230000011218 segmentation Effects 0.000 claims description 6
- 230000008030 elimination Effects 0.000 claims description 5
- 238000003379 elimination reaction Methods 0.000 claims description 5
- 239000000284 extract Substances 0.000 claims description 5
- 238000013507 mapping Methods 0.000 claims description 5
- 238000007781 pre-processing Methods 0.000 claims description 4
- 230000000717 retained effect Effects 0.000 claims description 3
- 238000010276 construction Methods 0.000 claims description 2
- 238000006467 substitution reaction Methods 0.000 claims description 2
- 238000002474 experimental method Methods 0.000 claims 1
- 230000014616 translation Effects 0.000 abstract 4
- 230000001131 transforming effect Effects 0.000 abstract 1
- 239000011159 matrix material Substances 0.000 description 12
- 238000004364 calculation method Methods 0.000 description 10
- 230000001537 neural effect Effects 0.000 description 5
- 238000013075 data extraction Methods 0.000 description 3
- 238000006243 chemical reaction Methods 0.000 description 2
- 210000002569 neuron Anatomy 0.000 description 2
- 238000010200 validation analysis Methods 0.000 description 2
- 244000068485 Convallaria majalis Species 0.000 description 1
- 230000009286 beneficial effect Effects 0.000 description 1
- 238000004140 cleaning Methods 0.000 description 1
- 238000013527 convolutional neural network Methods 0.000 description 1
- 238000012937 correction Methods 0.000 description 1
- 230000007123 defense Effects 0.000 description 1
- 230000007812 deficiency Effects 0.000 description 1
- 238000011161 development Methods 0.000 description 1
- 230000018109 developmental process Effects 0.000 description 1
- 238000010586 diagram Methods 0.000 description 1
- 230000008034 disappearance Effects 0.000 description 1
- 230000000694 effects Effects 0.000 description 1
- 238000005516 engineering process Methods 0.000 description 1
- 238000011156 evaluation Methods 0.000 description 1
- 238000004880 explosion Methods 0.000 description 1
- 239000000203 mixture Substances 0.000 description 1
- 230000000306 recurrent effect Effects 0.000 description 1
- 238000011160 research Methods 0.000 description 1
- 230000001960 triggered effect Effects 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F40/00—Handling natural language data
- G06F40/40—Processing or translation of natural language
- G06F40/58—Use of machine translation, e.g. for multi-lingual retrieval, for server-side translation for client devices or for real-time translation
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F16/00—Information retrieval; Database structures therefor; File system structures therefor
- G06F16/90—Details of database functions independent of the retrieved data types
- G06F16/903—Querying
- G06F16/90335—Query processing
- G06F16/90344—Query processing by using string matching techniques
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F40/00—Handling natural language data
- G06F40/20—Natural language analysis
- G06F40/279—Recognition of textual entities
- G06F40/289—Phrasal analysis, e.g. finite state techniques or chunking
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
- G06N3/044—Recurrent networks, e.g. Hopfield networks
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
- G06N3/045—Combinations of networks
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
- G06N3/047—Probabilistic or stochastic networks
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/08—Learning methods
- G06N3/084—Backpropagation, e.g. using gradient descent
Definitions
- the invention relates to the field of machine translation, in particular to a method for translating root nouns in English, French and German into Chinese based on a Transformer deep learning model.
- Neural machine translation usually adopts an encoder-decoder framework to achieve end-to-end translation between natural languages, and the Transformer model is the best of many neural machine translation models.
- the most significant difference between the Transformer model and other neural machine translation models is that the model completely relies on the attention mechanism, abandoning the recurrent neural network and convolutional neural network used in the traditional neural machine translation model, which makes the Transformer model largely alleviated
- the problem of gradient disappearance and gradient explosion improves the parallel computing ability of the model and shortens the training time of the model.
- the object of the present invention is to aim at the limitations and deficiencies of existing translation systems in the process of translating foreign language place names into Chinese, to provide a method for translating the roots of multilingual place names based on Transformer model, so as to obtain efficient and reasonable translation of place names in English, French and German. Chinese translation result.
- the present invention is achieved through the following steps in order to solve the above-mentioned problems:
- Step 1 First, preprocess the original foreign language place name corpus and the corresponding Chinese translation corpus;
- Step 2 Then, based on the knowledge base of place names and language rules composed of the collected and sorted rules of place names in various languages and language features, and combined with the language features of foreign language place names, identify the language of the input foreign language place name;
- Step 3 According to the language information of the recognized foreign language place names, select the place name root extraction rule corresponding to the language from the place name root extraction database, extract the root part of the foreign language place name, and use the Chinese place name root extraction rule to extract the place name corresponding to the Chinese translation. root part;
- Step 4 Convert the foreign language place name and the root text corresponding to the Chinese translation into a character set, and use the one-hot encoding and the character embedding model constructed by the shallow feedforward neural network to obtain the corresponding character vector of each foreign language character and Chinese character;
- Step 5 Train and fine-tune the Transformer model, and adjust the output dimension of the word embedding layer, the number of encoder layers, the number of self-attention mechanisms, the output dimension of the feedforward neural network, and the batch size based on the BLEU (Bilingual Evaluation Understudy) score.
- the number of processing, the number of pre-training, and the values of the seven hyperparameters of the drop regularization probability enable the Transformer model to achieve the highest BLEU score in the translation results of the test set;
- Step 6 Extract the root part of the place name to be translated into Chinese according to steps 1, 2 and 3, and convert the extraction result into a character vector and input it into the trained and fine-tuned Transformer model, and output the corresponding root Chinese translation result.
- the above-mentioned preprocessing includes removal of special characters of place names, expansion of abbreviations of foreign language place names, unified lowercase processing of foreign language place names, and replacement of pronunciation symbols.
- the present invention constructs a knowledge base of language rules for basic place names by summarizing and summarizing the words that appear frequently in English, French and German place names and can clearly distinguish the three languages.
- the present invention can be further expanded in combination with common names and place names in English, French and German summarized in a third-party knowledge base, and a place name language rules knowledge base is established to assist the language identification of place names.
- the above-mentioned place name root extraction includes the elimination of the common names of place names and the words that play a turning role in place names, that is, by constructing a place name elimination thesaurus, the generalized vocabulary of foreign and Chinese place names and the words that play a turning role are stored in it. After the preprocessed foreign and Chinese place names are processed by word segmentation, each word segmentation result is compared with the place name elimination lexicon through the index, and only the unmatched word segmentation results are retained, so as to obtain the root of the foreign language and Chinese place names.
- the conversion of the extraction result into a character vector is to convert the toponym root character represented by the one-hot encoding into a character vector by constructing a shallow feedforward neural network.
- the above-mentioned fine-tuning Transformer model is to determine the word embedding layer output dimension, the number of encoder layers, the number of self-attention mechanisms, the output dimension of the feed-forward neural network, the number of batches, the number of pre-training times, and the drop regularization by controlling variables. Probability of local optimal values for seven hyperparameters.
- the BLEU score of the model with different values of the hyperparameter on the test set is evaluated, so as to determine the hyperparameter The best value within the range of values.
- the above-mentioned model training times are not less than 50,000.
- the present invention has the following beneficial technical effects:
- the present invention focuses on the end-to-end translation between the noun root of a foreign language and a root of a Chinese place name, realizes the extraction of the noun root in the foreign language place name and the Chinese place name through the method based on the knowledge base, and converts the foreign language place name and the Chinese place name through the character embedding model.
- the result of word root extraction is further transformed into a character set, which is used as the input of the Transformer model in the form of a special character set, which expands the contextual dependence of the place name sequence, so as to obtain a better translation result of the place name root.
- the present invention summarizes and sorts out the foreign language features, corresponding language place name features and person name features, converts the above features into corresponding rules, and builds a knowledge base of place name language rules. Using the constructed knowledge base of place name language rules to identify the input foreign language place name language, thereby reducing the dependence on manual work.
- the present invention summarizes and sorts out the components of the foreign language place names involved, classifies each component, converts the occurrence rules into rules, and constructs a place noun root extraction rule base.
- the constructed toponymic root extraction rule base is used to extract the root part of the input foreign language toponyms, thereby significantly improving the translation efficiency of toponymic roots.
- Fig. 1 is the flow chart of the Chinese translation method of foreign language geographical nouns of the present invention
- Fig. 2 is the flow chart of obtaining root character vector of the present invention
- Fig. 3 is the Transformer model architecture diagram that the present invention relates to
- FIG. 4 is a flow chart of the calculation of the multi-head attention mechanism in the Transformer model involved in the present invention.
- the Chinese translation method of multilingual geographical nouns based on the Transformer deep learning model includes the following steps:
- the foreign language place name corpus is uniformly processed in lowercase and replaced by diacritics. For example, "New York” and “new york”, “cafe” and “café” both point to the same place name.
- the character replacement method unifies the format of the foreign language place name corpus.
- Training and fine-tuning the Transformer model is a Chinese translation model of foreign geographical nouns, and the training corpus is shown in Table 1.
- the data required for model training consists of the foreign language gazetteer root and the corresponding Chinese gazetteer root dataset divided into training set, validation set and test set according to the ratio of 7:2:1.
- the training set is the data required for model training
- the validation set is the data set used by the model to judge the performance of the model after training a fixed number of times, which can effectively indicate whether the model is in a state of overfitting or underfitting
- the test set is to judge whether the model is trained or not.
- the main body of the Transformer model is composed of an encoder (Encoder) and a decoder (Decoder).
- the input of the encoder and the decoder are the character vector of foreign language place names and the corresponding Chinese place name character vector respectively, and the dimension of the character vector is composed of words
- the output dimension of the embedding layer is controlled.
- one-step position encoding processing is performed, and a matrix M pe of the same dimension is added to each character vector.
- the calculation formula is:
- EncoderInput V ei (V ci )+M pe
- the self-attention mechanism is triggered, and the character vector will be multiplied by the matrices W q , W k , W v respectively to obtain the query matrix Q , the key matrix K and the value matrix V, the output Z calculation formula of the self-attention mechanism is:
- MultiHead(Z1,Z2,...,Zn) Concat(Z1,Z2,...,Zn)W o
- the model Before the output of the multi-head self-attention mechanism enters the feedforward neural network, the model performs a residual connection operation on it, combining the input information of the encoder with the output of the multi-head self-attention mechanism.
- the specific calculation formula is:
- LayerNorm is a regularization operation.
- Z1, Z2, ..., Zn are used as the input of the feedforward neural network, and the output dimension of the feedforward neural network is controlled by the output dimension of the feedforward neural network.
- the output of the feedforward neural network also needs a residual connection and regularization operation before it can be input into the next coding layer. In this residual connection and regularization operation, the output of the feedforward neural network needs to be the same as the first residual. Addition of Z1, Z2, ..., Zn after difference join and regularization operation.
- the operations performed in each coding layer after that are consistent with the above operations, and the number of coding layers is controlled by the number of encoder/decoder layers.
- the operation in the encoder is roughly the same as the decoder, except that the input to the decoder is a character vector of the root character set of Chinese toponyms, and the encoder-decoder attention is added in each decoding layer compared to the encoding layer
- the force mechanism combines the matrix output from the decoder and the output of the multi-head attention mechanism obtained in the encoding layer, fusing the input and output latent features.
- the Transformer model builds a feedforward neural network layer and a softmax layer to operate on the output of the encoder.
- the feedforward neural network layer maps the output of the encoder to a vector with the same dimension as the dictionary, and the softmax layer converts the mapped vector is the probability, and the character corresponding to the maximum probability is used as the output, and the final output of the model is composed of each output character.
- the batch size determines the amount of data after the training data is divided into batches
- the number of pre-training determines the number of times the model is pre-trained before formal training
- the probability of discarding regularization determines that all neurons in the model training process do not update parameter neurons proportion.
- the dynamic composition method of geographic model network service mainly consists of the following three parts:
- place name root data extraction module The preprocessing result of the place name source data "hazardville fire department" and the corresponding Chinese translation “Hazardville Fire Department" are used as the input of the place name root data extraction module.
- the place name root data extraction module first extracts the place name root part according to the place name splitting rules.
- the toponymic roots extracted from the input placenames are "hazardville” and "Hazardville”.
- the place name splitting rules are summarized after analyzing the characteristics of English and Chinese place names. Among them, the English place name splitting rules will filter out place name prefixes, place name suffixes and place name special words. As shown in Table 1, place name prefix words mainly include location.
- place name suffix words mainly include three categories: natural environment generic name, administrative division generic name and point of interest generic name; place name special words are a collection of words that play a turning or successor role in the word order in place names, while Chinese place names are split. rules such as
- place-name prefixes and place-name suffixes are filtered out, and the content of Chinese place-name prefixes and place-name suffixes is similar to that of English place-name prefixes and place-name suffixes.
- the toponymic root data is first converted into a character set, and then the shallow neural network constructed by the word-embedding layer in the open source PyTorch converts the geographical name data in the form of characters into a vector form that can be understood by computers.
- the process of vectorization of “hazardville” through shallow neural network is shown in Figure 2.
- Table 3 Examples of corpora required for Transformer model training and fine-tuning
- the method for obtaining the local optimal value of the other 6 parameters in the Transformer model including the input dimension, the output dimension of the feedforward layer, the number of coding layers and the number of batches, is the same as the above method.
- the input of the encoder and decoder in the Transformer model is the character vector of the English place name character set and the corresponding Chinese translation place name character set, respectively.
- the specific architecture of the model is shown in Figure 3.
- the character vector will be input to the encoder and decoder.
- For position encoding processing add a matrix M pe of the same dimension to each character vector V ci in the character set.
- the calculation formula is:
- the position-encoded character vector is input to the encoder and decoder, it is multiplied by the matrices W q , W k , W v to obtain the query matrix Q, the key matrix K and the value matrix V, and the calculation formula of the output Z of the self-attention mechanism for:
- d k is the dimension of the character vector
- the output of the multi-head self-attention mechanism is to connect the outputs of all self-attention mechanisms together and multiply it by the matrix W o , and the calculation formula is:
- MultiHead(Z1,Z2,...,Zn) Concat(Z1,Z2,...,Zn)W o
- the model Before the output of the multi-head self-attention mechanism enters the feedforward neural network, the model performs a residual connection operation on it, combining the input information of the encoder or decoder with the output of the multi-head self-attention mechanism.
- the calculation formula is:
- LayerNorm is a regularization operation, Z1, Z2, ..., Zn after residual connection and regularization operation are used as the input of the feedforward neural network, so as to model the potential mapping relationship between the source language and the target language.
- the output of the feedforward neural network needs a residual connection and regularization operation before it can be input to the next encoding layer or decoding layer.
- the output of the feedforward neural network needs to be The first residual connection and Z1, Z2, ..., Zn are added after the regularization operation.
- the specific calculation flow of the multi-head attention mechanism is shown in Figure 4.
- the operation in the decoder is roughly the same as the encoder, the difference is that the encoder-decoder attention mechanism is added in each decoding layer compared with the encoding layer, and the matrix output from the encoder and the multi-head attention obtained in the decoding layer are combined. Force mechanism outputs are combined, fusing input and output latent features.
- the Transformer model builds a feedforward neural network layer and a softmax layer to operate on the output of the decoder.
- the feedforward neural network layer maps the output of the decoder to a vector with the same dimension as the dictionary, and the softmax layer converts the mapped vector is the probability, and the character corresponding to the maximum probability is used as the output.
- the final output of the model is composed of each output character. Combined with this example, the final output of the model is "Hazzardville".
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Physics & Mathematics (AREA)
- Computational Linguistics (AREA)
- General Physics & Mathematics (AREA)
- General Engineering & Computer Science (AREA)
- Artificial Intelligence (AREA)
- General Health & Medical Sciences (AREA)
- Health & Medical Sciences (AREA)
- Data Mining & Analysis (AREA)
- Computing Systems (AREA)
- Evolutionary Computation (AREA)
- Molecular Biology (AREA)
- Biophysics (AREA)
- Life Sciences & Earth Sciences (AREA)
- Biomedical Technology (AREA)
- Mathematical Physics (AREA)
- Software Systems (AREA)
- Audiology, Speech & Language Pathology (AREA)
- Databases & Information Systems (AREA)
- Probability & Statistics with Applications (AREA)
- Machine Translation (AREA)
Abstract
Description
地名作为不可或缺的基础地理信息和社会公共信息,是各类社会信息关联的重要桥梁,在国家和社会管理、经济发展、文化建设、国防外交等方面发挥着重要作用。经济交往过程中大量外文地名的出现急需提出一种能合理地翻译外文地名的方法。
英语原文 | 标准翻译参照 |
Union | 尤宁 |
Pelham | 佩勒姆 |
Saul | 萨于勒 |
Donhead | 唐黑德 |
St Mary | 圣玛丽 |
Tuttington | 塔廷顿 |
Mayflower | 梅弗劳尔 |
Macclesfield | 麦克尔斯菲尔德 |
Claims (10)
- 一种基于Transformer深度学习模型的多语种地名词根汉译方法,其特征在于包括以下步骤:步骤1:对原始外文地名语料和对应中文翻译语料进行预处理;步骤2:基于由收集、整理的各语种地名、语言特征获取的规则所组成的地名语种规则知识库并结合外文地名的语种特征,识别输入外文地名的语种;步骤3:根据识别出的外文地名的语种信息,从地名词根抽取库中选择与语种相对应的地名词根抽取规则,提取外文地名的词根部分,利用中文地名词根抽取规则抽取对应中文翻译中的地名词根部分;步骤4:将外文地名和对应中文翻译的词根文本转化为字符集合,并利用独热编码与由浅层前馈神经网络构建的字符嵌入模型获取每个外文字符和中文字符相应的字符向量;步骤5:训练和微调Transformer模型,以BLEU得分为依据来调整词嵌入层输出维度、编码器层数、自注意力机制数、前馈神经网络输出维度、批处理数量、预训练次数和丢弃正则化概率七个超参数的取值,使得Transformer模型对测试集的翻译结果能取得最高的BLEU得分;步骤6:按照步骤1、2和3提取待汉译地名的词根部分,并将提取结果转化为字符向量输入到训练、微调完毕的Transformer模型中,输出相应的词根汉译结果。
- 根据权利要求1所述的基于Transformer深度学习模型的多语种地名词根汉译方法,其特征在于,所述预处理包括地名特殊字符剔除处理、外文地名缩写部分扩充处理和外文地名统一小写化处理和发音符号替代处理。
- 根据权利要求2所述的基于Transformer深度学习模型的多语种地名词根汉译方法,其特征在于,构建特殊字符库、缩写-全称映射库和发音符号替换映射库,并以上述知识库为基础,以遍历地名字符串的方式实现所述地名特殊字符剔除处理、外文地名缩写部分扩充处理和外文地名统一小写化处理。
- 根据权利要求1所述的基于Transformer深度学习模型的多语种地名词根汉译方法,其特征在于,通过归纳总结获得英语、法语和德语地名中出现频率高且能清晰区分三种语言的单词构建基础地名语种规则知识库。
- 根据权利要求4所述的基于Transformer深度学习模型的多语种地名词根汉译方法,其特征在于,基于所述基础地名语种规则知识库可结合第三方知识库中归纳的英语、法语和德语中常用人名、地名做进一步扩充,建立地名语种规则知识库辅助地名的语种识别。
- 根据权利要求1所述的基于Transformer深度学习模型的多语种地名词根汉译方法,其特征在于,所述地名词根提取包含对地名通名和地名中起到转折作用词汇的剔除,即通过构 建一个地名剔除词库,将归纳整理的外文、中文地名常用通名词汇和起到转折作用的词汇储存其中,预处理后的外文、中文地名经过分词处理后,将每个分词结果通过索引与地名剔除词库对比,仅保留不能匹配的分词结果,从而获得外文、中文地名的词根。
- 根据权利要求1所述的基于Transformer深度学习模型的多语种地名词根汉译方法,其特征在于,步骤6中所述将提取结果转化为字符向量是通过构建浅层前馈神经网络将由独热编码表示的地名词根字符转化为字符向量。
- 根据权利要求1所述的基于Transformer深度学习模型的多语种地名词根汉译方法,其特征在于所述微调Transformer模型是通过控制变量的方法设置对照实验来确定词嵌入层输出维度、编码器层数、自注意力机制数、前馈神经网络输出维度、批处理数量、预训练次数和丢弃正则化概率七个超参数的局部最优取值。
- 根据权利要求8所述的基于Transformer深度学习模型的多语种地名词根汉译方法,其特征在于通过采用固定其他超参数不变,改变上述七个超参数中某个超参数的取值,经过模型训练后评价该超参数的不同取值模型在测试集上的BLEU得分,从而判定该超参数在取值范围内的最优取值。
- 根据权利要求9所述的基于Transformer深度学习模型的多语种地名词根汉译方法,其特征在于所述模型训练次数不低于50000。
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
JP2021528844A JP2022552029A (ja) | 2020-09-15 | 2020-12-14 | Transformerのディープラーニングモデルに基づいて多言語による地名の語根を中国語に翻訳する方法 |
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202010967634.XA CN112084796B (zh) | 2020-09-15 | 2020-09-15 | 一种基于Transformer深度学习模型的多语种地名词根汉译方法 |
CN202010967634.X | 2020-09-15 |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2022057116A1 true WO2022057116A1 (zh) | 2022-03-24 |
Family
ID=73737117
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/CN2020/136009 WO2022057116A1 (zh) | 2020-09-15 | 2020-12-14 | 一种基于Transformer深度学习模型的多语种地名词根汉译方法 |
Country Status (3)
Country | Link |
---|---|
JP (1) | JP2022552029A (zh) |
CN (1) | CN112084796B (zh) |
WO (1) | WO2022057116A1 (zh) |
Cited By (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN112101348A (zh) * | 2020-08-28 | 2020-12-18 | 广州探迹科技有限公司 | 多语种端到端ocr算法及系统 |
CN114626363A (zh) * | 2022-05-16 | 2022-06-14 | 天津大学 | 一种基于翻译的跨语言短语结构分析方法及装置 |
CN114821257A (zh) * | 2022-04-26 | 2022-07-29 | 中国科学院大学 | 导航中视频流与自然语言的智能处理方法和装置、设备 |
CN114897004A (zh) * | 2022-04-15 | 2022-08-12 | 成都理工大学 | 一种基于深度学习Transformer模型的梯形堆积核脉冲识别方法 |
CN117592462A (zh) * | 2024-01-18 | 2024-02-23 | 航天宏图信息技术股份有限公司 | 基于地物群的开源地名数据的相关性处理方法及装置 |
Families Citing this family (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN113239707A (zh) * | 2021-03-01 | 2021-08-10 | 北京小米移动软件有限公司 | 文本翻译方法、文本翻译装置及存储介质 |
CN113393445B (zh) * | 2021-06-21 | 2022-08-23 | 上海交通大学医学院附属新华医院 | 乳腺癌影像确定方法及系统 |
AU2021104429A4 (en) * | 2021-07-22 | 2021-09-16 | Chinese Academy Of Surveying And Mapping | Machine Translation Method for French Geographical Names |
Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20140163951A1 (en) * | 2012-12-07 | 2014-06-12 | Xerox Corporation | Hybrid adaptation of named entity recognition |
CN104331401A (zh) * | 2014-11-25 | 2015-02-04 | 中国农业银行股份有限公司 | 一种翻译方法及系统 |
CN111222342A (zh) * | 2020-04-15 | 2020-06-02 | 北京金山数字娱乐科技有限公司 | 一种翻译方法和装置 |
CN111310456A (zh) * | 2020-02-13 | 2020-06-19 | 支付宝(杭州)信息技术有限公司 | 一种实体名称匹配方法、装置及设备 |
Family Cites Families (12)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JPH0754522B2 (ja) * | 1987-02-27 | 1995-06-07 | 日本電信電話株式会社 | 自然言語複合名詞解析・変換方式 |
JP6144458B2 (ja) * | 2012-03-06 | 2017-06-07 | 日本放送協会 | 手話翻訳装置及び手話翻訳プログラム |
CN108563640A (zh) * | 2018-04-24 | 2018-09-21 | 中译语通科技股份有限公司 | 一种多语言对的神经网络机器翻译方法及系统 |
CN109829173B (zh) * | 2019-01-21 | 2023-09-29 | 中国测绘科学研究院 | 一种英文地名翻译方法及装置 |
CN109902312B (zh) * | 2019-03-01 | 2023-07-11 | 北京金山数字娱乐科技有限公司 | 一种翻译方法及装置、翻译模型的训练方法及装置 |
CN110457715B (zh) * | 2019-07-15 | 2022-12-13 | 昆明理工大学 | 融入分类词典的汉越神经机器翻译集外词处理方法 |
CN111008517A (zh) * | 2019-10-30 | 2020-04-14 | 天津大学 | 一种基于张量分解技术的神经语言模型的压缩方法 |
CN111178091B (zh) * | 2019-12-20 | 2023-05-09 | 沈阳雅译网络技术有限公司 | 一种多维度的中英双语数据清洗方法 |
CN111209749A (zh) * | 2020-01-02 | 2020-05-29 | 湖北大学 | 一种将深度学习应用于中文分词的方法 |
CN111368035A (zh) * | 2020-03-03 | 2020-07-03 | 新疆大学 | 一种基于神经网络的汉维-维汉机构名词典的挖掘系统 |
CN111444343B (zh) * | 2020-03-24 | 2021-04-06 | 昆明理工大学 | 基于知识表示的跨境民族文化文本分类方法 |
CN111581988B (zh) * | 2020-05-09 | 2022-04-29 | 浙江大学 | 一种基于任务层面课程式学习的非自回归机器翻译模型的训练方法和训练系统 |
-
2020
- 2020-09-15 CN CN202010967634.XA patent/CN112084796B/zh active Active
- 2020-12-14 WO PCT/CN2020/136009 patent/WO2022057116A1/zh active Application Filing
- 2020-12-14 JP JP2021528844A patent/JP2022552029A/ja active Pending
Patent Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20140163951A1 (en) * | 2012-12-07 | 2014-06-12 | Xerox Corporation | Hybrid adaptation of named entity recognition |
CN104331401A (zh) * | 2014-11-25 | 2015-02-04 | 中国农业银行股份有限公司 | 一种翻译方法及系统 |
CN111310456A (zh) * | 2020-02-13 | 2020-06-19 | 支付宝(杭州)信息技术有限公司 | 一种实体名称匹配方法、装置及设备 |
CN111222342A (zh) * | 2020-04-15 | 2020-06-02 | 北京金山数字娱乐科技有限公司 | 一种翻译方法和装置 |
Non-Patent Citations (1)
Title |
---|
AN, SUYALA: "Chinese-Mongolian Organization Name Translation Based on Transformer", JOURNAL OF CHINESE INFORMATION PROCESSING, vol. 34, no. 1, 1 January 2020 (2020-01-01), pages 58 - 62, XP055912721, ISSN: 1003-0077 * |
Cited By (7)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN112101348A (zh) * | 2020-08-28 | 2020-12-18 | 广州探迹科技有限公司 | 多语种端到端ocr算法及系统 |
CN114897004A (zh) * | 2022-04-15 | 2022-08-12 | 成都理工大学 | 一种基于深度学习Transformer模型的梯形堆积核脉冲识别方法 |
CN114821257A (zh) * | 2022-04-26 | 2022-07-29 | 中国科学院大学 | 导航中视频流与自然语言的智能处理方法和装置、设备 |
CN114821257B (zh) * | 2022-04-26 | 2024-04-05 | 中国科学院大学 | 导航中视频流与自然语言的智能处理方法和装置、设备 |
CN114626363A (zh) * | 2022-05-16 | 2022-06-14 | 天津大学 | 一种基于翻译的跨语言短语结构分析方法及装置 |
CN117592462A (zh) * | 2024-01-18 | 2024-02-23 | 航天宏图信息技术股份有限公司 | 基于地物群的开源地名数据的相关性处理方法及装置 |
CN117592462B (zh) * | 2024-01-18 | 2024-04-16 | 航天宏图信息技术股份有限公司 | 基于地物群的开源地名数据的相关性处理方法及装置 |
Also Published As
Publication number | Publication date |
---|---|
CN112084796A (zh) | 2020-12-15 |
JP2022552029A (ja) | 2022-12-15 |
CN112084796B (zh) | 2021-04-09 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
WO2022057116A1 (zh) | 一种基于Transformer深度学习模型的多语种地名词根汉译方法 | |
CN111382580B (zh) | 一种面向神经机器翻译的编码器-解码器框架预训练方法 | |
CN110543639B (zh) | 一种基于预训练Transformer语言模型的英文句子简化算法 | |
CN111767718B (zh) | 一种基于弱化语法错误特征表示的中文语法错误更正方法 | |
US11573957B2 (en) | Natural language processing engine for translating questions into executable database queries | |
CN101788978B (zh) | 一种拼音和汉字相结合的汉外口语自动翻译方法 | |
CN112925918B (zh) | 一种基于疾病领域知识图谱的问答匹配系统 | |
Jian et al. | [Retracted] LSTM‐Based Attentional Embedding for English Machine Translation | |
JP2020190970A (ja) | 文書処理装置およびその方法、プログラム | |
CN114780582A (zh) | 基于表格问答的自然答案生成系统及其方法 | |
CN114429132A (zh) | 一种基于混合格自注意力网络的命名实体识别方法和装置 | |
CN110502759B (zh) | 融入分类词典的汉越混合网络神经机器翻译集外词处理方法 | |
CN113408307B (zh) | 一种基于翻译模板的神经机器翻译方法 | |
Abandah et al. | Correcting arabic soft spelling mistakes using bilstm-based machine learning | |
CN112257460B (zh) | 基于枢轴的汉越联合训练神经机器翻译方法 | |
Vykhovanets et al. | An overview of phonetic encoding algorithms | |
Smadja et al. | Translating collocations for use in bilingual lexicons | |
Kuo et al. | A phonetic similarity model for automatic extraction of transliteration pairs | |
Mi et al. | A neural network based model for loanword identification in Uyghur | |
CN113705223A (zh) | 以读者为中心的个性化英文文本简化方法 | |
CN114185573A (zh) | 一种人机交互机器翻译系统的实现和在线更新系统及方法 | |
JP3825645B2 (ja) | 表現変換方法及び表現変換装置 | |
CN115329784B (zh) | 基于预训练模型的句子复述生成系统 | |
CN111046665A (zh) | 一种领域术语语义漂移抽取方法 | |
Seresangtakul et al. | Thai-Isarn dialect parallel corpus construction for machine translation |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
ENP | Entry into the national phase |
Ref document number: 2021528844 Country of ref document: JP Kind code of ref document: A |
|
121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 20953971 Country of ref document: EP Kind code of ref document: A1 |
|
NENP | Non-entry into the national phase |
Ref country code: DE |
|
122 | Ep: pct application non-entry in european phase |
Ref document number: 20953971 Country of ref document: EP Kind code of ref document: A1 |
|
122 | Ep: pct application non-entry in european phase |
Ref document number: 20953971 Country of ref document: EP Kind code of ref document: A1 |
|
32PN | Ep: public notification in the ep bulletin as address of the adressee cannot be established |
Free format text: NOTING OF LOSS OF RIGHTS PURSUANT TO RULE 112(1) EPC (EPO FORM 1205A DATED 12.09.2023) |