WO2019041524A1 - Procédé, appareil électronique et support de stockage lisible par ordinateur permettant de générer une étiquette de grappe - Google Patents
Procédé, appareil électronique et support de stockage lisible par ordinateur permettant de générer une étiquette de grappe Download PDFInfo
- Publication number
- WO2019041524A1 WO2019041524A1 PCT/CN2017/108807 CN2017108807W WO2019041524A1 WO 2019041524 A1 WO2019041524 A1 WO 2019041524A1 CN 2017108807 W CN2017108807 W CN 2017108807W WO 2019041524 A1 WO2019041524 A1 WO 2019041524A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- cluster
- keywords
- keyword
- calculation formula
- extracting
- Prior art date
Links
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F16/00—Information retrieval; Database structures therefor; File system structures therefor
- G06F16/30—Information retrieval; Database structures therefor; File system structures therefor of unstructured textual data
- G06F16/36—Creation of semantic tools, e.g. ontology or thesauri
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F16/00—Information retrieval; Database structures therefor; File system structures therefor
- G06F16/30—Information retrieval; Database structures therefor; File system structures therefor of unstructured textual data
- G06F16/35—Clustering; Classification
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/23—Clustering techniques
- G06F18/232—Non-hierarchical techniques
Definitions
- the present application relates to the field of computer information technology, and in particular, to a cluster label generation method, an electronic device, and a computer readable storage medium.
- the present application proposes a cluster label generation method, an electronic device, and a computer readable storage medium, and optimizes a clustering keyword extraction process on a semantic level by a preset naive Bayesian calculation formula, and The label extraction of clustered text is optimized.
- the present application provides an electronic device including a memory and a processor, on which is stored a cluster label generation system executable on the processor, the The class tag generation system is implemented by the processor to implement the following steps:
- the most discriminating keywords are extracted from the keywords of each cluster, and are recorded as labels of each cluster.
- the extracting the representative keywords comprises: extracting keywords of each cluster according to the conditional probability value size of the words.
- the extracting representative keywords includes:
- conditional probability values of each word calculated by each cluster are sorted in descending order, and a preset number of keywords are extracted and recorded as cluster keywords.
- the extracting the most distinguishing keyword comprises: extracting the most discriminating keyword from the keywords of each cluster according to the transition probability value between the words and the preset naive Bayes calculation formula.
- the extracting the most distinguishing keywords comprises:
- conditional probability values of each keyword recalculated for each cluster are sorted in descending order, and the keyword with the highest conditional probability value is extracted and recorded as a clustering label.
- the preset naive Bayesian calculation formula is set to formula 1:
- Equation 1 S represents a piece of text consisting of n words W1, W2, ... Wn, and Wi represents a word in the semantic network relationship constructed by the piece of text;
- the preset transition probability calculation formula is set to Equation 2:
- Wi) Pt(Wj
- Equation 2 m represents the number of clusters after text clustering, t represents one of the clusters, Wi and Wj represent keywords extracted by each cluster, and Pt(Wj
- the present application further provides a cluster label generation method, which is applied to an electronic device, and the method includes:
- the most discriminating keywords are extracted from the keywords of each cluster, and are recorded as labels of each cluster.
- the extracting the representative keywords includes: extracting keywords of each cluster according to the conditional probability value size of the words, specifically including:
- conditional probability values of each word calculated by each cluster are sorted in descending order, and a preset number of keywords are extracted and recorded as cluster keywords.
- the extracting the most distinguishing keyword comprises: extracting the most distinctive keyword from each cluster of keywords according to a transition probability value between words and a preset naive Bayesian calculation formula, Specifically include:
- conditional probability values of each keyword recalculated for each cluster are sorted in descending order, and the keyword with the highest conditional probability value is extracted and recorded as a clustering label.
- the present application further provides a computer readable storage medium storing a cluster label generation system, the cluster label generation system
- the step of causing the at least one processor to perform the clustering label generation method as described above may be performed by at least one processor.
- the electronic device, the cluster label generation method and the computer readable storage medium proposed by the present application optimize the extraction of clustering keywords on the semantic level by using a preset naive Bayesian calculation formula. process. Further, the label extraction of the clustered text is also optimized, so that the extracted cluster labels have high discrimination and recognition.
- 1 is a schematic diagram of an optional hardware architecture of an electronic device of the present application
- FIG. 2 is a schematic diagram of a program module of an embodiment of a cluster label generation system in an electronic device of the present application
- FIG. 3 is a schematic diagram of an implementation process of an embodiment of a method for generating a cluster label according to the present application.
- FIG. 1 it is a schematic diagram of an optional hardware architecture of the electronic device 2 of the present application.
- the electronic device 2 may include, but is not limited to, a memory 21, a processor 22, and a network interface 23 that can communicate with each other through a system bus. It is pointed out that FIG. 1 only shows the electronic device 2 with the components 21-23, but it should be understood that not all illustrated components are required to be implemented, and more or fewer components may be implemented instead.
- the electronic device 2 may be a computing device such as a rack server, a blade server, a tower server, or a rack server.
- the electronic device 2 may be an independent server or a server cluster composed of multiple servers. .
- the memory 21 includes at least one type of readable storage medium including a flash memory, a hard disk, a multimedia card, a card type memory (eg, SD or DX memory, etc.), a random access memory (RAM), a static Random access memory (SRAM), read only memory (ROM), electrically erasable programmable read only memory (EEPROM), programmable read only memory (PROM), magnetic memory, magnetic disk, optical disk, and the like.
- the memory 21 may be an internal storage unit of the electronic device 2, such as a hard disk or memory of the electronic device 2.
- the memory 21 may also be an external storage device of the electronic device 2, such as a plug-in hard disk equipped on the electronic device 2, a smart memory card (SMC), and a secure digital device. (Secure Digital, SD) card, flash card, etc.
- the memory 21 may also include both an internal storage unit of the electronic device 2 and an external storage device thereof.
- the memory 21 is generally used to store an operating system installed in the electronic device 2 and various types of application software, such as program codes of the cluster tag generation system 20. Further, the memory 21 can also be used to temporarily store various types of data that have been output or are to be output.
- the processor 22 may be a Central Processing Unit (CPU), controller, microcontroller, microprocessor, or other data processing chip in some embodiments.
- the processor 22 is typically used to control the overall operation of the electronic device 2, such as performing control and processing related to data interaction or communication with the electronic device 2.
- the processor 22 is configured to run program code or process data stored in the memory 21, for example, to run the cluster.
- the tag generation system 20 and the like.
- the network interface 23 may comprise a wireless network interface or a wired network interface, which is typically used to establish a communication connection between the electronic device 2 and other electronic devices.
- the network interface 23 is configured to connect the electronic device 2 to an external data platform through a network, and establish a data transmission channel and a communication connection between the electronic device 2 and an external data platform.
- the network may be an intranet, an Internet, a Global System of Mobile communication (GSM), a Wideband Code Division Multiple Access (WCDMA), a 4G network, or a 5G network.
- Wireless or wired networks such as network, Bluetooth, Wi-Fi, etc.
- the cluster tag generation system 20 may be divided into one or more program modules, the one or more program modules being stored in the memory 21 and being processed by one or more processors. (Processing in the present embodiment for the processor 22) to complete the application.
- the cluster tag generation system 20 can be divided into a construction module 201, an extraction module 202, and a generation module 203.
- the program module referred to in the present application refers to a series of computer program instruction segments capable of performing a specific function, and is more suitable than the program to describe the execution process of the cluster tag generation system 20 in the electronic device 2. The function of each program module 201-203 will be described in detail below.
- the building module 201 is configured to construct a semantic network relationship between words in each cluster for the text clustering result.
- the text clustering is performed for the unsupervised corpus
- the clustering method may adopt the Text-rank clustering algorithm
- the text clustering result may be the text summary information or the like.
- the semantic network relationship is used to describe the concept and state of an object and its relationship. It consists of an arc between a node and a node. The node represents a concept (event, thing, etc.), and the arc represents the relationship between concepts.
- the extracting module 202 is configured to extract a representative keyword from the semantic network relationship constructed by each cluster, and record it as a clustering keyword.
- the extracting the representative keywords comprises: extracting keywords of each cluster according to the conditional probability value size of the words.
- S represents a piece of text
- Wi represents a word in the semantic network relationship constructed by the piece of text
- P the conditional probability value of each word in the semantic network relationship constructed by each cluster
- the conditional probability values of each word calculated by each cluster are sorted in descending order, and a preset number (for example, three) of keywords is extracted and recorded as a clustering keyword.
- the clustering keyword is a word that best represents the semantics of the piece of text.
- the conditional probability value is calculated according to a preset naive Bayesian Drawn.
- the preset naive Bayesian calculation formula can be set as the following formula 1 (LaTex version).
- Equation 1 can also be expressed as follows:
- Equation 1 represents the probability that the text S appears in the case where the given word Wi appears, the right half of the equation is the product calculation formula, and n represents the number of words in the text S.
- the generating module 203 is configured to extract, from the keywords of each cluster, the most distinctive keywords, and record the labels of each cluster.
- the extracting the most distinguishing keyword comprises: according to the transition probability value between the words and the preset naive Bayes calculation formula, from each cluster of keywords Extract the most discriminating keywords. Specifically, first, according to a preset transition probability calculation formula, a transition probability value between the keywords in the total document aggregated by all the documents of each cluster is calculated.
- the preset transition probability calculation formula may be set to the following formula 2.
- Wi) Representation The transition probability of the keywords Wi to Wj in the total document in which all the documents of the t-th cluster are aggregated.
- the formula for calculating the transition probability between keywords in the first cluster is:
- Wi) P1(Wj
- transition probability value between the keywords in each cluster is substituted into the preset naive Bayesian calculation formula (the above formula 1), and the conditional probability value of each keyword is recalculated (final result) Is a multiplication of a transfer matrix).
- the conditional probability values of each keyword recalculated for each cluster are sorted in descending order, and the keyword with the highest conditional probability value is extracted and recorded as a clustering label.
- the recalculated conditional probability value represents the discriminative level of each keyword. The higher the conditional probability value recalculated by a keyword, the higher the discriminativeness, and the more suitable for clustering labels.
- multiple keywords with higher discrimination may be selected from the keywords of each cluster, as each cluster. label.
- the cluster label generation system 20 optimizes the extraction process of cluster keywords on the semantic level by using a preset naive Bayesian calculation formula. Further, the label extraction of the clustered text is also optimized, so that the extracted cluster labels have high discrimination and recognition.
- the present application also proposes a cluster label generation method.
- FIG. 3 it is a schematic flowchart of an implementation process of an embodiment of a method for generating a cluster label of the present application.
- the order of execution of the steps in the flowchart shown in FIG. 3 may be changed according to different requirements, and some steps may be omitted.
- Step S31 constructing a semantic network relationship between words in each cluster for the text clustering result.
- the text clustering is performed for the unsupervised corpus
- the clustering method may adopt the Text-rank clustering algorithm
- the text clustering result may be the text summary information or the like.
- the semantic network relationship is used to describe the concept and state of an object and its relationship. It consists of an arc between a node and a node. The node represents a concept (event, thing, etc.), and the arc represents the relationship between concepts.
- step S32 a representative keyword is extracted from the semantic network relationship constructed by each cluster, and is recorded as a clustering keyword.
- the extracting the representative keywords comprises: extracting keywords of each cluster according to the conditional probability value size of the words.
- S represents a piece of text
- Wi represents a word in the semantic network relationship constructed by the piece of text
- P the conditional probability value of each word in the semantic network relationship constructed by each cluster
- the conditional probability values of each word calculated by each cluster are sorted in descending order, and a preset number (for example, three) of keywords is extracted and recorded as a clustering keyword.
- the clustering keyword is a word that best represents the semantics of the piece of text.
- the conditional probability value is obtained according to a preset naive Bayesian calculation formula.
- the preset naive Bayesian calculation formula can be set as the following formula 1 (LaTex version).
- Equation 1 can also be expressed as follows:
- Equation 1 represents the probability that the text S appears in the case where the given word Wi appears, the right half of the equation is the product calculation formula, and n represents the number of words in the text S.
- step S33 the most discriminating keywords are extracted from the keywords of each cluster, and are recorded as labels of each cluster.
- the extracting the most distinguishing keyword comprises: according to the transition probability value between the words and the preset naive Bayes calculation formula, from each cluster of keywords Extract the most discriminating keywords. Specifically, first, according to a preset transition probability calculation formula, a transition probability value between the keywords in the total document aggregated by all the documents of each cluster is calculated.
- the preset transition probability calculation formula may be set to the following formula 2.
- m represents the number of clusters after text clustering
- t represents one of the clusters (eg, the first cluster)
- Wi and Wj represent keywords extracted by each cluster
- the formula for calculating the transition probability between keywords in the first cluster is:
- Wi) P1(Wj
- transition probability value between the keywords in each cluster is substituted into the preset naive Bayesian calculation formula (the above formula 1), and the conditional probability value of each keyword is recalculated (final result) Is a multiplication of a transfer matrix).
- the conditional probability values of each keyword recalculated for each cluster are sorted in descending order, and the keyword with the highest conditional probability value is extracted and recorded as a clustering label.
- the recalculated conditional probability value represents the discriminative level of each keyword. The higher the conditional probability value recalculated by a keyword, the higher the discriminativeness, and the more suitable for clustering labels.
- multiple keywords with higher discrimination may be selected from the keywords of each cluster, as each cluster. label.
- the cluster label generation method proposed by the present application optimizes the extraction process of the clustering keywords on the semantic level by the preset naive Bayesian calculation formula. Further, the label extraction of the clustered text is also optimized, so that the extracted cluster labels have high discrimination and recognition.
- the present application further provides a computer readable storage medium (such as a ROM/RAM, a magnetic disk, an optical disk), where the computer readable storage medium stores a cluster label generation system 20, the aggregation
- the class tag generation system 20 can be executed by at least one processor 22 to cause the at least one processor 22 to perform the steps of the cluster tag generation method as described above.
- the foregoing embodiment method can be implemented by means of software plus a necessary general hardware platform, and can also be implemented by hardware, but in many cases, the former is A better implementation.
- the technical solution of the present application which is essential or contributes to the prior art, may be embodied in the form of a software product stored in a storage medium (such as ROM/RAM, disk,
- the optical disc includes a number of instructions for causing a terminal device (which may be a mobile phone, a computer, a server, an air conditioner, or a network device, etc.) to perform the methods described in various embodiments of the present application.
Abstract
L'invention concerne un procédé permettant de générer une étiquette de grappe, ledit procédé consistant à : construire, pour des résultats de regroupement de texte, une relation de réseau sémantique pour des mots dans chaque grappe (S31); extraire des mots-clés représentatifs à partir de la relation de réseau sémantique construite à partir de chacune des grappes, puis marquer ceux-ci en tant que mots-clés de grappe (S32); et extraire le mot-clé le plus discriminant à partir des mots-clés de chaque grappe et marquer celui-ci en tant qu'étiquette de ladite grappe (S33). L'invention améliore ainsi la capacité de discrimination et d'identification d'une étiquette d'une grappe.
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201710776351.5 | 2017-08-31 | ||
CN201710776351.5A CN107679084B (zh) | 2017-08-31 | 2017-08-31 | 聚类标签生成方法、电子设备及计算机可读存储介质 |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2019041524A1 true WO2019041524A1 (fr) | 2019-03-07 |
Family
ID=61134852
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/CN2017/108807 WO2019041524A1 (fr) | 2017-08-31 | 2017-10-31 | Procédé, appareil électronique et support de stockage lisible par ordinateur permettant de générer une étiquette de grappe |
Country Status (2)
Country | Link |
---|---|
CN (1) | CN107679084B (fr) |
WO (1) | WO2019041524A1 (fr) |
Cited By (7)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN110647914A (zh) * | 2019-08-14 | 2020-01-03 | 深圳壹账通智能科技有限公司 | 智能服务水平训练方法、装置及计算机可读存储介质 |
CN111814475A (zh) * | 2019-04-09 | 2020-10-23 | Oppo广东移动通信有限公司 | 用户画像构建方法、装置、存储介质和电子设备 |
CN112487132A (zh) * | 2019-09-12 | 2021-03-12 | 北京国双科技有限公司 | 关键词的确定方法和相关设备 |
CN112560455A (zh) * | 2019-09-26 | 2021-03-26 | 北京国双科技有限公司 | 一种数据的融合方法及相关系统 |
CN112579769A (zh) * | 2019-09-30 | 2021-03-30 | 北京国双科技有限公司 | 关键词的聚类方法、装置、存储介质和电子设备 |
CN112784063A (zh) * | 2019-03-15 | 2021-05-11 | 北京金山数字娱乐科技有限公司 | 一种成语知识图谱构建方法及装置 |
CN113204653A (zh) * | 2021-06-04 | 2021-08-03 | 中国银行股份有限公司 | 需求价值的标注方法、装置、计算机设备及可读存储介质 |
Families Citing this family (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN113688209B (zh) * | 2021-09-01 | 2023-08-25 | 江苏省城市规划设计研究院有限公司 | 一种通过调节关键词依存关系的文本语义网络构建方法 |
CN115618085B (zh) * | 2022-10-21 | 2024-04-05 | 华信咨询设计研究院有限公司 | 一种基于动态标签的接口数据暴露探测方法 |
Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20070078832A1 (en) * | 2005-09-30 | 2007-04-05 | Yahoo! Inc. | Method and system for using smart tags and a recommendation engine using smart tags |
CN103164463A (zh) * | 2011-12-16 | 2013-06-19 | 国际商业机器公司 | 推荐标签的方法和装置 |
CN105677769A (zh) * | 2015-12-29 | 2016-06-15 | 广州神马移动信息科技有限公司 | 一种基于潜在狄利克雷分配(lda)模型的关键词推荐方法和系统 |
CN106844748A (zh) * | 2017-02-16 | 2017-06-13 | 湖北文理学院 | 文本聚类方法、装置及电子设备 |
Family Cites Families (11)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JPH0934861A (ja) * | 1995-07-17 | 1997-02-07 | Olympus Optical Co Ltd | クラスタ分類装置 |
CN102207945B (zh) * | 2010-05-11 | 2013-10-23 | 天津海量信息技术有限公司 | 基于知识网络的文本标引系统及其方法 |
CN102207946B (zh) * | 2010-06-29 | 2013-10-23 | 天津海量信息技术有限公司 | 一种知识网络的半自动生成方法 |
CN102708096B (zh) * | 2012-05-29 | 2014-10-15 | 代松 | 一种基于语义的网络智能舆情监测系统及其工作方法 |
CN104239300B (zh) * | 2013-06-06 | 2017-10-20 | 富士通株式会社 | 从文本中挖掘语义关键词的方法和设备 |
CN103544255B (zh) * | 2013-10-15 | 2017-01-11 | 常州大学 | 基于文本语义相关的网络舆情信息分析方法 |
US10191978B2 (en) * | 2014-01-03 | 2019-01-29 | Verint Systems Ltd. | Labeling/naming of themes |
CN104794500A (zh) * | 2015-05-11 | 2015-07-22 | 苏州大学 | 一种tri-training半监督学习方法及装置 |
CN104951432B (zh) * | 2015-05-21 | 2019-01-11 | 腾讯科技(深圳)有限公司 | 一种对信息进行处理的方法及装置 |
CN105468713B (zh) * | 2015-11-19 | 2018-07-17 | 西安交通大学 | 一种多模型融合的短文本分类方法 |
CN106940726B (zh) * | 2017-03-22 | 2020-09-01 | 山东大学 | 一种基于知识网络的创意自动生成方法与终端 |
-
2017
- 2017-08-31 CN CN201710776351.5A patent/CN107679084B/zh active Active
- 2017-10-31 WO PCT/CN2017/108807 patent/WO2019041524A1/fr active Application Filing
Patent Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20070078832A1 (en) * | 2005-09-30 | 2007-04-05 | Yahoo! Inc. | Method and system for using smart tags and a recommendation engine using smart tags |
CN103164463A (zh) * | 2011-12-16 | 2013-06-19 | 国际商业机器公司 | 推荐标签的方法和装置 |
CN105677769A (zh) * | 2015-12-29 | 2016-06-15 | 广州神马移动信息科技有限公司 | 一种基于潜在狄利克雷分配(lda)模型的关键词推荐方法和系统 |
CN106844748A (zh) * | 2017-02-16 | 2017-06-13 | 湖北文理学院 | 文本聚类方法、装置及电子设备 |
Cited By (7)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN112784063A (zh) * | 2019-03-15 | 2021-05-11 | 北京金山数字娱乐科技有限公司 | 一种成语知识图谱构建方法及装置 |
CN111814475A (zh) * | 2019-04-09 | 2020-10-23 | Oppo广东移动通信有限公司 | 用户画像构建方法、装置、存储介质和电子设备 |
CN110647914A (zh) * | 2019-08-14 | 2020-01-03 | 深圳壹账通智能科技有限公司 | 智能服务水平训练方法、装置及计算机可读存储介质 |
CN112487132A (zh) * | 2019-09-12 | 2021-03-12 | 北京国双科技有限公司 | 关键词的确定方法和相关设备 |
CN112560455A (zh) * | 2019-09-26 | 2021-03-26 | 北京国双科技有限公司 | 一种数据的融合方法及相关系统 |
CN112579769A (zh) * | 2019-09-30 | 2021-03-30 | 北京国双科技有限公司 | 关键词的聚类方法、装置、存储介质和电子设备 |
CN113204653A (zh) * | 2021-06-04 | 2021-08-03 | 中国银行股份有限公司 | 需求价值的标注方法、装置、计算机设备及可读存储介质 |
Also Published As
Publication number | Publication date |
---|---|
CN107679084B (zh) | 2021-09-28 |
CN107679084A (zh) | 2018-02-09 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
WO2019041524A1 (fr) | Procédé, appareil électronique et support de stockage lisible par ordinateur permettant de générer une étiquette de grappe | |
WO2021068610A1 (fr) | Procédé et appareil de recommandation de ressources, dispositif électronique et support d'informations | |
WO2019062021A1 (fr) | Procédé de distribution sélective d'une publicité de prêt dans un programme d'application, dispositif électronique, et support | |
WO2019085120A1 (fr) | Procédé de recommandation à filtrage collaboratif, dispositif électronique, et support d'informations lisible par ordinateur | |
WO2019061991A1 (fr) | Procédé de modélisation de plateforme de modèle universel à éléments multiples, dispositif électronique, et support d'informations lisible par ordinateur | |
CN110598001A (zh) | 联合实体关系抽取方法、装置及存储介质 | |
WO2020000717A1 (fr) | Procédé et dispositif de classification de pages web et support de stockage lisible par ordinateur | |
US10650274B2 (en) | Image clustering method, image clustering system, and image clustering server | |
CN109189888B (zh) | 电子装置、侵权分析的方法及存储介质 | |
WO2016015621A1 (fr) | Procédé et système de reconnaissance de nom d'image de visage humain | |
US20220012231A1 (en) | Automatic content-based append detection | |
WO2019075967A1 (fr) | Procédé de reconnaissance de nom d'entreprise, dispositif électronique, et support d'informations lisible par ordinateur | |
WO2021051869A1 (fr) | Procédé d'arrangement d'agencement de données de texte, dispositif, appareil informatique et support de stockage | |
US9971809B1 (en) | Systems and methods for searching unstructured documents for structured data | |
WO2019062081A1 (fr) | Procédé de formation de profil de vendeur, dispositif électronique et support d'informations lisible par ordinateur | |
WO2020253043A1 (fr) | Procédé et appareil de classification de texte intelligents et support de stockage lisible par ordinateur | |
CN112632278A (zh) | 一种基于多标签分类的标注方法、装置、设备及存储介质 | |
US9355091B2 (en) | Systems and methods for language classification | |
US9330075B2 (en) | Method and apparatus for identifying garbage template article | |
WO2019041528A1 (fr) | Procédé, appareil électronique et support d'informations lisible par ordinateur permettant de déterminer la polarité d'un sentiment portant sur les actualités | |
CN110704608A (zh) | 文本主题生成方法、装置和计算机设备 | |
CN110781955A (zh) | 无标签对象的分类和检测套码的方法、装置及计算机可读存储介质 | |
WO2019119635A1 (fr) | Procédé de développement d'utilisateur initial, dispositif électronique et support de stockage lisible par ordinateur | |
US10331789B2 (en) | Semantic analysis apparatus, method, and non-transitory computer readable storage medium thereof | |
WO2019024231A1 (fr) | Procédé d'appariement automatique de données, dispositif électronique et support de stockage lisible par ordinateur |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
NENP | Non-entry into the national phase |
Ref country code: DE |
|
121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 17923073 Country of ref document: EP Kind code of ref document: A1 |
|
32PN | Ep: public notification in the ep bulletin as address of the adressee cannot be established |
Free format text: NOTING OF LOSS OF RIGHTS PURSUANT TO RULE 112(1) EPC (EPO FORM 1205A DATED 29.09.2020) |
|
122 | Ep: pct application non-entry in european phase |
Ref document number: 17923073 Country of ref document: EP Kind code of ref document: A1 |