WO2022088444A1 - 一种面向多任务语言模型的元-知识微调方法及平台 - Google Patents
一种面向多任务语言模型的元-知识微调方法及平台 Download PDFInfo
- Publication number
- WO2022088444A1 WO2022088444A1 PCT/CN2020/138014 CN2020138014W WO2022088444A1 WO 2022088444 A1 WO2022088444 A1 WO 2022088444A1 CN 2020138014 W CN2020138014 W CN 2020138014W WO 2022088444 A1 WO2022088444 A1 WO 2022088444A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- task
- meta
- model
- knowledge
- fine
- Prior art date
Links
- 238000000034 method Methods 0.000 title claims abstract description 24
- 230000006835 compression Effects 0.000 claims abstract description 38
- 238000007906 compression Methods 0.000 claims abstract description 38
- 238000012549 training Methods 0.000 claims abstract description 33
- 230000006870 function Effects 0.000 claims description 22
- 238000011524 similarity measure Methods 0.000 claims description 5
- 238000011176 pooling Methods 0.000 claims description 4
- 238000003058 natural language processing Methods 0.000 claims description 3
- 230000000694 effects Effects 0.000 abstract description 8
- 230000008094 contradictory effect Effects 0.000 description 2
- 238000013461 design Methods 0.000 description 2
- 230000007935 neutral effect Effects 0.000 description 2
- 230000003190 augmentative effect Effects 0.000 description 1
- 230000009286 beneficial effect Effects 0.000 description 1
- 238000004364 calculation method Methods 0.000 description 1
- 230000000052 comparative effect Effects 0.000 description 1
- 230000007812 deficiency Effects 0.000 description 1
- 238000010586 diagram Methods 0.000 description 1
- 230000008451 emotion Effects 0.000 description 1
- 238000012545 processing Methods 0.000 description 1
- 230000003014 reinforcing effect Effects 0.000 description 1
- 230000008685 targeting Effects 0.000 description 1
- 238000012360 testing method Methods 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F16/00—Information retrieval; Database structures therefor; File system structures therefor
- G06F16/30—Information retrieval; Database structures therefor; File system structures therefor of unstructured textual data
- G06F16/35—Clustering; Classification
- G06F16/355—Class or cluster creation or modification
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F40/00—Handling natural language data
- G06F40/30—Semantic analysis
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N5/00—Computing arrangements using knowledge-based models
- G06N5/04—Inference or reasoning models
Definitions
- the invention belongs to the field of language model compression, and in particular relates to a meta-knowledge fine-tuning method and platform for multi-task language models.
- the purpose of the present invention is to provide a meta-knowledge fine-tuning method and platform for multi-task language models in view of the deficiencies of the prior art.
- the present invention proposes cross-domain typicality score learning, uses this method to obtain highly transferable shared knowledge on different data sets of similar tasks, and introduces "meta-knowledge" to learn the process of learning similar tasks on different domains corresponding to different data sets Interconnect and reinforce each other, improve the fine-tuning effect of similar downstream tasks on different domain datasets in the application of language models in the smart city domain, and improve the parameter initialization and generalization capabilities of general language models for similar tasks.
- a multi-task language model-oriented meta-knowledge fine-tuning method including the following stages:
- the class prototypes of the cross-domain datasets of the same task from the datasets of the same task in different domains, centrally learn the embedding features of the prototypes of the corresponding domains of the task, and average the input texts of all the input texts of the same task in different domains. Embedding features, as the corresponding multi-domain class prototypes of the same type of task;
- the typicality score of the instance is calculated: d self represents the distance between the embedded feature of each instance and its own domain prototype, and d others represents the distance between the embedded feature of each instance and other domain prototypes; the typicality of each instance The score is defined as the linear combination of d self and d others ;
- the third stage, meta-knowledge fine-tuning network based on typicality score Using the typicality score obtained in the second stage as the weight coefficient of the meta-knowledge fine-tuning network, a multi-task typicality-sensitive label classification loss function is designed as the meta-knowledge fine-tuning Learn an objective function; this loss function penalizes the labels of instances of all domains where the text classifier mispredicts.
- M is the set of all class labels in the dataset; is the i-th instance in the k-th domain;
- ⁇ ( ) represents the output of the BERT model
- ⁇ is a predefined balance factor, 0 ⁇ 1; cos( ⁇ , ⁇ ) is the cosine similarity measure function; K is the number of domains; is the indicator function, if returns 1 if then returns 0, the index for summation; ⁇ m > 0 is weights of the same class The weights are the same.
- D represents the set of all domains; is the indicator function, if returns 1 if then return 0; Indicates prediction The probability that the class label of m is m; Embedding layer feature representing the token of "[CLS]" output by the last layer of the BERT model.
- a meta-knowledge fine-tuning platform for multi-task language models including the following components:
- Data loading component used to obtain training samples of a multi-task-oriented pre-trained language model, where the training samples are labeled text samples that satisfy supervised learning tasks;
- Automatic compression component used to automatically compress a multi-task-oriented pre-trained language model, including a pre-trained language model and a meta-knowledge fine-tuning module; wherein the meta-knowledge fine-tuning module is used for the pre-trained language generated by the automatic compression component
- the downstream task network is constructed on the model, and the meta-knowledge of the typical score is used to fine-tune the downstream task scene, and the final fine-tuned student model is output, that is, the pre-trained language model compression model containing the downstream tasks required by the landing user; the compressed model is output Go to the specified container for the login user to download, and present the comparison information of the model size before and after compression;
- the login user obtains the pre-trained language model compression model from the platform, and the user uses the compression model output by the automatic compression component to infer the new data of the natural language processing downstream task uploaded by the login user on the data set of the actual scene, and Presents comparison information of inference speed before and after compression.
- the present invention studies a multi-task language model-oriented meta-knowledge fine-tuning method based on cross-domain typicality score learning, the fine-tuning method of the downstream task-oriented pre-training language model is to fine-tune on the downstream task cross-domain data set,
- the effect of the compressed model obtained by fine-tuning is not limited to the specific dataset of this type of task.
- the downstream task is fine-tuned through the meta-knowledge fine-tuning network, thereby obtaining similar downstream tasks that are independent of the dataset. language model;
- the present invention proposes to learn highly transferable shared knowledge on different datasets of the same task, namely meta-knowledge; introducing meta-knowledge, and the meta-knowledge fine-tuning network combines the learning process on different domains corresponding to different datasets of the same task Interrelated and mutually reinforcing, improve the fine-tuning effect of similar downstream tasks on different domain datasets in the application of language models in the smart city domain, improve the parameter initialization and generalization capabilities of general language models for similar tasks, and finally obtain similar downstream task languages.
- Model
- the multi-task language model-oriented meta-knowledge fine-tuning platform of the present invention generates a general architecture for similar task language models, makes full use of the fine-tuned model architecture to improve the compression efficiency of downstream similar tasks, and can convert large-scale natural
- the language processing model is deployed on end-side devices such as small memory and limited resources, which promotes the implementation of general-purpose deep language models in the industry.
- FIG. 1 is an overall architecture diagram of the meta-knowledge fine-tuning method of the present invention.
- a meta-knowledge fine-tuning method and platform for multi-task language models of the present invention is based on cross-domain typicality score learning on the downstream task multi-domain data set of the pre-trained language model, using typical Fractional meta-knowledge fine-tunes downstream task scenarios, making it easier for meta-learners to fine-tune to any domain, and the learned knowledge is highly generalizable and transferable, rather than limited to a specific domain.
- the effect of the compression model is suitable for data scenarios of the same task and different domains.
- a meta-knowledge fine-tuning method for multi-task language model of the present invention specifically includes the following steps:
- Step 1 Calculate the class prototypes of the same task cross-domain datasets: Considering that the multi-domain class prototypes can summarize the key semantic features of the corresponding training datasets; therefore, from the datasets of different domains, centrally learn the prototypes of the corresponding domains of this type of task The embedding feature of the same type of task and multi-domain class prototype is generated. Specifically, for the BERT language model, the average embedding feature of all input texts in different domains of the same task is used as the class prototype corresponding to this type of task, where the average embedding feature is The output of the last layer of Transformer encoder corresponding to the current input instance is used to average the pooling layer output.
- the average embedding feature of all input texts in the kth domain Dk is taken as the class prototype corresponding to this domain.
- class prototype is the input BERT model
- the average pooling of the corresponding last layer Transformer encoder is calculated as follows:
- ⁇ ( ) means that the Embedding features that map to d-dimension.
- Step 2 Calculate the typicality score of the training instance: Considering that if the training instance is semantically close to the class prototype of its own domain, and is not too far from the class prototypes generated by other domains, the instance is considered to be typical and has a high High portability.
- the semantics of a training instance should include not only its associated features with its own domain, but also its associated features with other domains.
- a typical training instance is defined as a linear combination of the above two associated features. Specifically, d self is used to represent the distance between the embedded feature of each training instance and its own domain prototype, d others is used to represent the distance between the embedded feature of each training instance and other domain prototypes, and the typicality score of each training instance is defined as d self Linear combination with d others .
- the above single class prototype is further augmented to generate a certain category of class prototypes based on the clustering of multiple prototypes.
- the possible polarities include positive, negative, neutral and conflict.
- the generic class prototype corresponding to the category can be generated by clustering on multiple different datasets.
- the associated feature of each training instance with its own domain is that each training instance with its own domain prototype The cosine similarity measure distance of , i.e.
- the associated feature of each training instance with other domains is that each training instance Cosine similarity measure distance to class prototypes generated by other domains, i.e.
- Typical training example The characteristic score of :
- ⁇ is a predefined balance factor, 0 ⁇ 1, cos( ⁇ , ⁇ ) is the cosine similarity measure function, 1 ( ⁇ ) is the indicator function, and returns 1 if the input Boolean function is true, Otherwise, return 0.
- Step 3 Meta-knowledge fine-tuning network based on typicality score:
- the present invention proposes to design a multi-task typicality-sensitive label classification loss function based on cross-domain typical instance features. This loss function penalizes the labels of typical instances of all K domains where the text classifier mispredicts.
- the typicality score obtained in the second stage is used as the weight coefficient of the meta-knowledge fine-tuning network.
- the learning objective function of the meta-knowledge fine-tuning network is defined as:
- L T is a multi-task typicality-sensitive label classification loss function that penalizes the labels of typical instances of all K domains where the text classifier mispredicts.
- is the weight of each training instance. is the predicted instance
- the probability of the class label of m ⁇ M, the embedding layer of the token of the d-dimensional "[CLS]" of the last layer of BERT is used as the feature, and the express.
- the present invention is a multi-task language model-oriented meta-knowledge fine-tuning platform, comprising the following components:
- Data loading component used to obtain training samples of a multi-task-oriented pre-trained language model, where the training samples are labeled text samples that satisfy supervised learning tasks.
- Automatic compression component used to automatically compress multi-task-oriented pre-trained language models, including pre-trained language models and meta-knowledge fine-tuning modules.
- the meta-knowledge fine-tuning module is to build a downstream task network on the pre-trained language model generated by the automatic compression component, use the meta-knowledge of typical scores to fine-tune the downstream task scene, and output the final fine-tuned student model, that is, the login user
- a pre-trained language model compression model that includes downstream tasks as required; output the compressed model to a designated container for download by the logged-in user, and present a comparison of the size of the model before and after compression on the output compression model page of the platform information.
- the logged-in user obtains the pre-trained compression model from the platform, and the user uses the compression model output by the automatic compression component to infer the new data of the natural language processing downstream task uploaded by the logged-in user on the dataset of the actual scene; and The comparison information of the inference speed before and after compression is presented on the compression model inference page of the platform.
- the natural language inference task is to give a pair of sentences and determine whether the two sentences are semantically similar, contradictory, or neutral. Since it is also a classification problem, it is also called a sentence pair classification problem.
- the MNLI dataset provides training examples from multiple domains in order to infer whether two sentences are similar, contradictory, or unrelated.
- the BERT pre-training model generated by the automatic compression component is used to construct a model for the natural language inference task on the generated pre-training model; based on the student model obtained from the meta-knowledge fine-tuning module of the automatic compression component, fine-tuning is performed, and the pre-training language
- a downstream task network is constructed, and the meta-knowledge of the typical score is used to fine-tune the downstream task scene, and the final fine-tuned student model is output, that is, the pre-trained language model compression model containing the natural language inference task required by the landing user;
- the compressed model is output to a designated container for download by the logged-in user, and 5%, 10%, and 20% of the data in each domain are randomly sampled from the training
- Table 1 BERT model element for natural language inference task - comparative information before and after knowledge fine-tuning
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- General Physics & Mathematics (AREA)
- Physics & Mathematics (AREA)
- General Engineering & Computer Science (AREA)
- Computational Linguistics (AREA)
- Artificial Intelligence (AREA)
- Data Mining & Analysis (AREA)
- Databases & Information Systems (AREA)
- Evolutionary Computation (AREA)
- Computing Systems (AREA)
- Mathematical Physics (AREA)
- Software Systems (AREA)
- Health & Medical Sciences (AREA)
- Audiology, Speech & Language Pathology (AREA)
- General Health & Medical Sciences (AREA)
- Machine Translation (AREA)
Abstract
Description
方法 | 动物 | 植物 | 车辆 | 平均 |
元-知识微调前 | 93.6% | 91.8% | 84.2% | 89.3% |
元-知识微调后 | 94.5% | 92.3% | 90.2% | 92.3% |
Claims (5)
- 一种面向多任务语言模型的元-知识微调方法,其特征在于,包括以下几个阶段:第一阶段,计算同类任务跨域数据集的类原型:从同一类任务的不同域的数据集中,集中学习该类任务对应域的原型的嵌入特征,将同类任务不同域的所有输入文本的平均嵌入特征,作为对应的同一类任务多域的类原型;第二阶段,计算实例的典型性分数:采用d self表示每个实例的嵌入特征与自身域原型的距离,d others表示每个实例的嵌入特征与其它域原型的距离;每个实例的典型性分数定义为d self与d others的线性组合;第三阶段,基于典型性分数的元-知识微调网络:利用第二阶段得到的典型性分数作为元-知识微调网络的权重系数,设计多任务典型性敏感标签分类损失函数作为元-知识微调的学习目标函数;该损失函数惩罚语言模型预测错误的所有域的实例的标签。
- 一种基于权利要求1-4任一项所述面向多任务语言模型的元-知识微调方法的平台,其特征在于,包括以下组件:数据加载组件:用于获取面向多任务的预训练语言模型的训练样本,所述训练样本是满足监督学习任务的有标签的文本样本;自动压缩组件:用于将面向多任务的预训练语言模型自动压缩,包括预训练语言模型和元-知识微调模块;其中,所述元-知识微调模块用于在自动压缩组件生成的预训练语言模型上构建下游任务网络,利用典型性分数的元-知识对下游任务场景进行微调,输出最终微调好的学生模型,即登陆用户需求的包含下游任务的预训练语言模型压缩模型;将压缩模型输出到指定的容器,供登陆用户下载,并呈现压缩前后模型大小的对比信息;推理组件:登陆用户从平台获取预训练语言模型压缩模型,用户利用所述自动压缩组件输出的压缩模型在实际场景的数据集上对登陆用户上传的自然语言处理下游任务的新数据进行推理,并呈现压缩前后推理速度的对比信息。
Priority Applications (3)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
GB2214177.4A GB2609768A (en) | 2020-11-02 | 2020-12-21 | Multi-task language model-oriented meta-knowledge fine tuning method and platform |
JP2022567027A JP7283836B2 (ja) | 2020-11-02 | 2020-12-21 | マルチタスク言語モデル向けのメタ知識微調整方法及びプラットフォーム |
US17/531,813 US11354499B2 (en) | 2020-11-02 | 2021-11-22 | Meta-knowledge fine tuning method and platform for multi-task language model |
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202011202867.7 | 2020-11-02 | ||
CN202011202867.7A CN112100383B (zh) | 2020-11-02 | 2020-11-02 | 一种面向多任务语言模型的元-知识微调方法及平台 |
Related Child Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US17/531,813 Continuation US11354499B2 (en) | 2020-11-02 | 2021-11-22 | Meta-knowledge fine tuning method and platform for multi-task language model |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2022088444A1 true WO2022088444A1 (zh) | 2022-05-05 |
Family
ID=73784520
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/CN2020/138014 WO2022088444A1 (zh) | 2020-11-02 | 2020-12-21 | 一种面向多任务语言模型的元-知识微调方法及平台 |
Country Status (2)
Country | Link |
---|---|
CN (1) | CN112100383B (zh) |
WO (1) | WO2022088444A1 (zh) |
Cited By (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN114647732A (zh) * | 2022-05-23 | 2022-06-21 | 之江实验室 | 一种面向弱监督文本分类系统、方法和装置 |
CN115859175A (zh) * | 2023-02-16 | 2023-03-28 | 合肥综合性国家科学中心人工智能研究院(安徽省人工智能实验室) | 基于跨模态生成式学习的液压减震器设备异常检测方法 |
CN117436236A (zh) * | 2023-09-27 | 2024-01-23 | 四川大学 | 一种基于大模型的工艺流程智能规划方法 |
CN117669737A (zh) * | 2023-12-20 | 2024-03-08 | 中科星图数字地球合肥有限公司 | 一种端到端地理行业大语言模型构建及使用方法 |
CN117708337A (zh) * | 2024-02-05 | 2024-03-15 | 杭州杰竞科技有限公司 | 一种面向复杂定域的人机交互方法和系统 |
Families Citing this family (7)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN112100383B (zh) * | 2020-11-02 | 2021-02-19 | 之江实验室 | 一种面向多任务语言模型的元-知识微调方法及平台 |
JP7283836B2 (ja) | 2020-11-02 | 2023-05-30 | 之江実験室 | マルチタスク言語モデル向けのメタ知識微調整方法及びプラットフォーム |
CN112364945B (zh) * | 2021-01-12 | 2021-04-16 | 之江实验室 | 一种基于域-不变特征的元-知识微调方法及平台 |
GB2608344A (en) | 2021-01-12 | 2022-12-28 | Zhejiang Lab | Domain-invariant feature-based meta-knowledge fine-tuning method and platform |
CN113032559B (zh) * | 2021-03-15 | 2023-04-28 | 新疆大学 | 一种用于低资源黏着性语言文本分类的语言模型微调方法 |
CN113987209B (zh) * | 2021-11-04 | 2024-05-24 | 浙江大学 | 基于知识指导前缀微调的自然语言处理方法、装置、计算设备和存储介质 |
CN114780722B (zh) * | 2022-03-31 | 2024-05-14 | 北京理工大学 | 一种结合领域通用型语言模型的领域泛化方法 |
Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US10607598B1 (en) * | 2019-04-05 | 2020-03-31 | Capital One Services, Llc | Determining input data for speech processing |
CN111814448A (zh) * | 2020-07-03 | 2020-10-23 | 苏州思必驰信息科技有限公司 | 预训练语言模型量化方法和装置 |
CN111832282A (zh) * | 2020-07-16 | 2020-10-27 | 平安科技(深圳)有限公司 | 融合外部知识的bert模型的微调方法、装置及计算机设备 |
CN112100383A (zh) * | 2020-11-02 | 2020-12-18 | 之江实验室 | 一种面向多任务语言模型的元-知识微调方法及平台 |
Family Cites Families (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN107767954A (zh) * | 2017-10-16 | 2018-03-06 | 中国科学院地理科学与资源研究所 | 一种基于空间贝叶斯网络的环境健康风险监测预警系统及方法 |
CN108830287A (zh) * | 2018-04-18 | 2018-11-16 | 哈尔滨理工大学 | 基于残差连接的Inception网络结合多层GRU的中文图像语义描述方法 |
CN110909145B (zh) * | 2019-11-29 | 2022-08-09 | 支付宝(杭州)信息技术有限公司 | 针对多任务模型的训练方法及装置 |
CN111310848B (zh) * | 2020-02-28 | 2022-06-28 | 支付宝(杭州)信息技术有限公司 | 多任务模型的训练方法及装置 |
CN111291166B (zh) * | 2020-05-09 | 2020-11-03 | 支付宝(杭州)信息技术有限公司 | 基于Bert的语言模型的训练方法及装置 |
-
2020
- 2020-11-02 CN CN202011202867.7A patent/CN112100383B/zh active Active
- 2020-12-21 WO PCT/CN2020/138014 patent/WO2022088444A1/zh active Application Filing
Patent Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US10607598B1 (en) * | 2019-04-05 | 2020-03-31 | Capital One Services, Llc | Determining input data for speech processing |
CN111814448A (zh) * | 2020-07-03 | 2020-10-23 | 苏州思必驰信息科技有限公司 | 预训练语言模型量化方法和装置 |
CN111832282A (zh) * | 2020-07-16 | 2020-10-27 | 平安科技(深圳)有限公司 | 融合外部知识的bert模型的微调方法、装置及计算机设备 |
CN112100383A (zh) * | 2020-11-02 | 2020-12-18 | 之江实验室 | 一种面向多任务语言模型的元-知识微调方法及平台 |
Cited By (9)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN114647732A (zh) * | 2022-05-23 | 2022-06-21 | 之江实验室 | 一种面向弱监督文本分类系统、方法和装置 |
CN114647732B (zh) * | 2022-05-23 | 2022-09-06 | 之江实验室 | 一种面向弱监督文本分类系统、方法和装置 |
CN115859175A (zh) * | 2023-02-16 | 2023-03-28 | 合肥综合性国家科学中心人工智能研究院(安徽省人工智能实验室) | 基于跨模态生成式学习的液压减震器设备异常检测方法 |
CN117436236A (zh) * | 2023-09-27 | 2024-01-23 | 四川大学 | 一种基于大模型的工艺流程智能规划方法 |
CN117436236B (zh) * | 2023-09-27 | 2024-05-17 | 四川大学 | 一种基于大模型的工艺流程智能规划方法 |
CN117669737A (zh) * | 2023-12-20 | 2024-03-08 | 中科星图数字地球合肥有限公司 | 一种端到端地理行业大语言模型构建及使用方法 |
CN117669737B (zh) * | 2023-12-20 | 2024-04-26 | 中科星图数字地球合肥有限公司 | 一种端到端地理行业大语言模型构建及使用方法 |
CN117708337A (zh) * | 2024-02-05 | 2024-03-15 | 杭州杰竞科技有限公司 | 一种面向复杂定域的人机交互方法和系统 |
CN117708337B (zh) * | 2024-02-05 | 2024-04-26 | 杭州杰竞科技有限公司 | 一种面向复杂定域的人机交互方法和系统 |
Also Published As
Publication number | Publication date |
---|---|
CN112100383B (zh) | 2021-02-19 |
CN112100383A (zh) | 2020-12-18 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
WO2022088444A1 (zh) | 一种面向多任务语言模型的元-知识微调方法及平台 | |
CN110427461B (zh) | 智能问答信息处理方法、电子设备及计算机可读存储介质 | |
CN114064918B (zh) | 一种多模态事件知识图谱构建方法 | |
CN109815336B (zh) | 一种文本聚合方法及系统 | |
Zhang et al. | Sentiment Classification Based on Piecewise Pooling Convolutional Neural Network. | |
CN113407660B (zh) | 非结构化文本事件抽取方法 | |
CN112256866B (zh) | 一种基于深度学习的文本细粒度情感分析算法 | |
CN109271516B (zh) | 一种知识图谱中实体类型分类方法及系统 | |
CN108416032A (zh) | 一种文本分类方法、装置及存储介质 | |
CN114818703B (zh) | 基于BERT语言模型和TextCNN模型的多意图识别方法及系统 | |
CN112732872B (zh) | 面向生物医学文本的基于主题注意机制的多标签分类方法 | |
CN113849653B (zh) | 一种文本分类方法及装置 | |
CN116089873A (zh) | 模型训练方法、数据分类分级方法、装置、设备及介质 | |
CN112925904A (zh) | 一种基于Tucker分解的轻量级文本分类方法 | |
CN114756678A (zh) | 一种未知意图文本的识别方法及装置 | |
CN113743079A (zh) | 一种基于共现实体交互图的文本相似度计算方法及装置 | |
US11354499B2 (en) | Meta-knowledge fine tuning method and platform for multi-task language model | |
CN117033961A (zh) | 一种上下文语境感知的多模态图文分类方法 | |
CN116306869A (zh) | 训练文本分类模型的方法、文本分类方法及对应装置 | |
CN117216617A (zh) | 文本分类模型训练方法、装置、计算机设备和存储介质 | |
CN113239143B (zh) | 融合电网故障案例库的输变电设备故障处理方法及系统 | |
WO2022151553A1 (zh) | 一种基于域-不变特征的元-知识微调方法及平台 | |
CN114969341A (zh) | 一种针对餐饮行业评论的细粒度情感分析方法及装置 | |
CN113869051A (zh) | 一种基于深度学习的命名实体识别方法 | |
CN113076741A (zh) | 一种基于多语言文本数据分析方法 |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 20959601 Country of ref document: EP Kind code of ref document: A1 |
|
ENP | Entry into the national phase |
Ref document number: 202214177 Country of ref document: GB Kind code of ref document: A Free format text: PCT FILING DATE = 20201221 |
|
ENP | Entry into the national phase |
Ref document number: 2022567027 Country of ref document: JP Kind code of ref document: A |
|
NENP | Non-entry into the national phase |
Ref country code: DE |
|
122 | Ep: pct application non-entry in european phase |
Ref document number: 20959601 Country of ref document: EP Kind code of ref document: A1 |
|
32PN | Ep: public notification in the ep bulletin as address of the adressee cannot be established |
Free format text: NOTING OF LOSS OF RIGHTS PURSUANT TO RULE 112(1) EPC (EPO FORM 1205A DATED 28.09.2023) |