CN110866121A - Knowledge graph construction method for power field - Google Patents
Knowledge graph construction method for power field Download PDFInfo
- Publication number
- CN110866121A CN110866121A CN201910917049.6A CN201910917049A CN110866121A CN 110866121 A CN110866121 A CN 110866121A CN 201910917049 A CN201910917049 A CN 201910917049A CN 110866121 A CN110866121 A CN 110866121A
- Authority
- CN
- China
- Prior art keywords
- relation
- entity
- vector
- relationship
- sentence
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Pending
Links
- 238000010276 construction Methods 0.000 title claims abstract description 13
- 238000000034 method Methods 0.000 claims abstract description 35
- 238000012549 training Methods 0.000 claims abstract description 33
- 239000013598 vector Substances 0.000 claims description 53
- 238000000605 extraction Methods 0.000 claims description 19
- 230000007246 mechanism Effects 0.000 claims description 14
- 238000012360 testing method Methods 0.000 claims description 14
- 238000013135 deep learning Methods 0.000 claims description 10
- 238000013507 mapping Methods 0.000 claims description 6
- 238000013528 artificial neural network Methods 0.000 claims description 5
- 238000004364 calculation method Methods 0.000 claims description 4
- 239000011159 matrix material Substances 0.000 claims description 4
- 238000003058 natural language processing Methods 0.000 claims description 4
- 230000009466 transformation Effects 0.000 claims description 4
- 230000008569 process Effects 0.000 claims description 3
- 238000012216 screening Methods 0.000 claims description 3
- 238000002372 labelling Methods 0.000 claims description 2
- 210000002569 neuron Anatomy 0.000 claims description 2
- 230000009467 reduction Effects 0.000 abstract description 6
- 230000002950 deficient Effects 0.000 description 2
- 238000005516 engineering process Methods 0.000 description 2
- 230000002457 bidirectional effect Effects 0.000 description 1
- 238000013527 convolutional neural network Methods 0.000 description 1
- 238000010586 diagram Methods 0.000 description 1
- 238000012986 modification Methods 0.000 description 1
- 230000004048 modification Effects 0.000 description 1
- 238000003062 neural network model Methods 0.000 description 1
- 230000008520 organization Effects 0.000 description 1
- 230000001105 regulatory effect Effects 0.000 description 1
- 238000000638 solvent extraction Methods 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F16/00—Information retrieval; Database structures therefor; File system structures therefor
- G06F16/30—Information retrieval; Database structures therefor; File system structures therefor of unstructured textual data
- G06F16/36—Creation of semantic tools, e.g. ontology or thesauri
- G06F16/367—Ontology
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
- G06N3/045—Combinations of networks
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
- G06N3/049—Temporal neural networks, e.g. delay elements, oscillating neurons or pulsed inputs
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/08—Learning methods
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06Q—INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES; SYSTEMS OR METHODS SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES, NOT OTHERWISE PROVIDED FOR
- G06Q50/00—Information and communication technology [ICT] specially adapted for implementation of business processes of specific business sectors, e.g. utilities or tourism
- G06Q50/06—Energy or water supply
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Physics & Mathematics (AREA)
- Health & Medical Sciences (AREA)
- General Physics & Mathematics (AREA)
- Computational Linguistics (AREA)
- General Health & Medical Sciences (AREA)
- Life Sciences & Earth Sciences (AREA)
- Data Mining & Analysis (AREA)
- General Engineering & Computer Science (AREA)
- Business, Economics & Management (AREA)
- Molecular Biology (AREA)
- Computing Systems (AREA)
- Evolutionary Computation (AREA)
- Artificial Intelligence (AREA)
- Mathematical Physics (AREA)
- Software Systems (AREA)
- Biomedical Technology (AREA)
- Biophysics (AREA)
- Economics (AREA)
- Water Supply & Treatment (AREA)
- Public Health (AREA)
- Human Resources & Organizations (AREA)
- Marketing (AREA)
- Primary Health Care (AREA)
- Strategic Management (AREA)
- Tourism & Hospitality (AREA)
- General Business, Economics & Management (AREA)
- Animal Behavior & Ethology (AREA)
- Databases & Information Systems (AREA)
- Machine Translation (AREA)
Abstract
The method constructs the knowledge graph facing the electric power field, solves the problem of noise reduction in the construction training set of remote supervised learning, improves the accuracy of entity identification, can construct an accurate knowledge graph in the electric power field, constructs an accurate knowledge base for the electric power field, provides a foundation for knowledge discovery in the electric power field, and lays a foundation for the extension of the knowledge graph in the electric power field.
Description
Technical Field
The invention belongs to the field of electric power planning, and particularly relates to a knowledge graph construction method for the electric power field.
Background
The knowledge graph aims to describe various entities or concepts existing in the real world and the relationship thereof, and forms a huge semantic network graph. A knowledge graph mainly describes entities existing in the real world and relationships among the entities, and two entities and the relationships among the entities form a triple of the knowledge graph. In the construction of the knowledge graph, the main tasks are named entity identification and entity relation extraction.
Named entity recognition, also known as entity extraction technology and entity partitioning technology, is a sub-field of natural language processing technology. The goal is to extract named entities referred to in unstructured text, including but not limited to, person names, organization names, place name medical terms, regulatory terms, time, quantity, monetary value, and the like. The prior art is well established in named entity recognition. At present, with the appearance of word vectors, a convolutional neural network model and a bidirectional long-short term memory network are applied to named entity recognition in combination with a conditional random field model, and high accuracy is obtained. The existing named entity recognition has high recognition rate on common texts, but in the professional field, due to the particularity of field vocabularies, the recognition accuracy on proper nouns in the field of some specific grammar structures is not high.
In the aspect of entity relationship extraction, the relationship extraction method of remote supervision learning draws attention of broad scholars. Because supervised learning is used on the premise that a large amount of artificial labeled corpora are needed, the deep learning method based on the neural network usually needs quite large labeled corpora to perform model training. In order to solve the problem of insufficient data in the supervised learning process, Mintz et al propose a remote supervised learning method, which utilizes the existing knowledge in a knowledge base to automatically generate a large amount of labeled data by aligning the knowledge with the text. The generated data is then used for training of a relational extraction neural network model.
Because the training set constructed based on the remote supervision method is too hard to assume, the generated training text is relatively noisy. In addition, if the knowledge base in the field is deficient, the method needs a lot of manual work to construct the initial corpus, which is difficult to be completed without experts in the field, while the power field is in the current situation of deficient knowledge base.
Disclosure of Invention
The invention aims to construct a knowledge graph facing the electric power field, and because the structured data of the electric power text is less, the relation between entities needs to be extracted from a large amount of unstructured data to construct the knowledge graph. The training set is required to be constructed through a remote supervision method, and the problem to be solved firstly is the noise reduction problem of the training set constructed through remote supervision and learning. In addition, in the aspect of named entity recognition, the accuracy of entity recognition is improved by adding a professional dictionary in the power field.
The invention provides a power field knowledge graph construction method, which at least comprises the following steps:
step 1, acquiring structured data and semi-structured and unstructured data in the power field on a network;
step 3, named entity recognition is carried out through a deep learning Chinese named entity recognition method;
and 4, extracting entity relations by adopting a remote supervised learning method, and realizing construction of the knowledge graph facing the power field.
The knowledge graph oriented to the electric power field is constructed, and because the structured data of the electric power text is less, the relationships among the entities need to be extracted from a large amount of unstructured data to construct the knowledge graph. And the problem to be solved is the noise reduction problem of the training set constructed by remote supervised learning. In addition, in the aspect of named entity recognition, the accuracy of entity recognition is improved by adding a professional dictionary in the power field. In conclusion, the method and the device solve the problem of noise reduction in the construction training set of the remote supervised learning, improve the accuracy of entity identification, construct an accurate knowledge map in the power field, construct an accurate knowledge base for the power field, provide a foundation for knowledge discovery in the power field, and lay a foundation for extension of the knowledge map in the power field.
Drawings
FIG. 1 is a BiLSTM-CRF model in named entity recognition.
FIG. 2 shows a sentence vector representation module in entity relationship extraction.
Fig. 3 is a diagram of a remote supervised learning extraction framework.
Detailed Description
For a better understanding of the present invention, the method and system of the present invention will be further described with reference to the following description of the embodiments in conjunction with the accompanying drawings.
In the following detailed description, numerous specific details are set forth in order to provide a thorough understanding of the invention. It will be understood by those skilled in the art, however, that the present invention may be practiced without these specific details. In the embodiments, well-known methods, procedures, components, and so forth have not been described in detail as not to unnecessarily obscure the embodiments.
The invention provides a power field knowledge graph construction method, which at least comprises the following steps:
step 1, acquiring structured data and semi-structured and unstructured data in the power field on a network;
step 3, named entity recognition is carried out through a deep learning Chinese named entity recognition method;
and 4, extracting entity relations by adopting a remote supervised learning method, and realizing construction of the knowledge graph facing the power field.
Preferably, in step 1, structured data and semi-structured and unstructured data in the power field are acquired on the network, and specifically encyclopedia texts in the power field are crawled on the principle of breadth first.
Preferably, in step 3, the named entity recognition is performed by a deep learning chinese named entity recognition method, which specifically includes:
step 3-1, carrying out distributed expression on the words;
step 3-2, carrying out model training by using a deep learning network under supervision;
and 3-3, labeling each word in the sequence by using the context information.
Preferably, in step 3-2, supervised model training is performed by using a deep learning network, which specifically includes:
taking a power field data set as a training corpus, and training a word vector through a Skip-gram mode; the training network adopts a three-layer neural network consisting of an input layer, a hidden layer and an output layer, wherein 100 neurons of the hidden layer are set according to experience; the neural network module adopts BilSTM, regards semantic association of words in sentences as a sequence problem, and stores historical information for learning during network training; and (4) considering context correlation, performing sequence annotation by adopting a CRF (domain gradient random access) model, and separating the correlation of an output level.
Preferably, in step 4, the entity relationship extraction is performed by using a remote supervised learning method, which specifically includes:
step 4-1, entity alignment is carried out, and a relation instance set for training and testing is constructed in an entity alignment mode;
mapping the triple relations in the knowledge base to a training document for entity alignment, and generating a relation instance set Q:
Q={qn|qn=(sm,ei,rk,ej),sm∈D} (1)
wherein e isi、ejIs two entities, rkAs a relationship of two entities in a knowledge base, smFor sentences of entity pairs in corpus D, qnIs the generated relationship instance;
and 4-2, performing intra-sentence relation extraction by adopting a relation extraction model based on an attention mechanism.
Preferably, in the step 4-1, entity alignment is performed, and a relationship instance set for training and testing is constructed in an entity alignment manner, specifically including:
step 4-1-1, a mapping step, namely mapping each entity into a sentence of a text, wherein the co-occurrence of a pair of entities in each sentence is used as a relation example, and a plurality of relation examples with the same relation form a relation system;
4-1-2, training, namely aligning entities by using triples extracted from the structured data and encyclopedic texts;
and 4-1-3, a testing step, namely generating candidate relation pairs in pairs by using all entities in the test set in a permutation and combination mode, and generating corresponding test examples and relation examples by using the candidate relation pairs and the test corpus in an entity alignment mode.
Preferably, in the step 4-2, the intra-sentence relation extraction is performed by using a relation extraction model based on an attention mechanism,
the attention-based relationship extraction model mainly comprises two parts: a sentence vector representation module and a sentence-level attention mechanism module;
the sentence vector representation module is used for obtaining the characteristic representation of each relation instance in the relation system;
and the sentence-level attention mechanism module is used for measuring the importance degree of each relation instance relative to the relation system.
Preferably, in the sentence vector representation module, word vector representation is performed by using word2vec method, and the relative relationship between the vocabulary and the entity in the sentence is captured by using the word position vector;
the word vector for the ith word in the sentence is denoted wiTo makeAndthe expression vocabulary wiWord position vectors from two entities, using tiAs a word wiIs expressed as shown in equation (2):
using BilSTM to obtain the forward state and backward state of each vocabulary, and converting the forward state of each vocabulary into forward stateAnd backward stateThe result of the concatenation is expressed as the state of the vocabulary, as shown in equation (3):
after obtaining the state information of all the words, the vector s of the sentenceiThe representation can be determined jointly by all the lexical states inside it:
preferably, wherein the sentence-level attention mechanism module comprises: the system comprises an attention mechanism calculation unit, an entity feature representation layer, a relation package feature representation layer, a hidden layer and an output layer;
the attention mechanism calculating unit is used for calculating weights of different instances in the relationship package to obtain vector representation of each relationship system. In the weight calculation process, three kinds of characteristic information of a concept vector, a sentence mark vector and a target relation vector are fused on the basis of a sentence vector, wherein the concept vector eiIncluding descriptor vector ciAnd a superior-inferior token vector qi;
The characteristic representation layer of the relationship system is used for obtaining the characteristics of the relationship system, and the characteristics are jointly determined by the relationship examples in the package; the relation system S is composed of n relation examples, S ═ S1,s2,...,snAnd f, the feature vector u of the relationship packet S can be obtained by formula (5):
wherein, αkIs the weight of the kth relationship instance, skIs the feature vector of the kth relationship instance;
the entity feature representation layer is used for obtaining the abstract features of the entity by using the BilSTM, and specifically comprises the following steps:
the respective probability vectors e of the entities in the relation1And e2Taken together, BilSTM is used to obtain the forward state of each entityAnd the state of the consequentThe forward state and the back term state of the entity are merged as shown in equation (6).
After obtaining the respective state vectors of the entities in the relationship, summing the state vectors as the final feature representation e of the entity pairf;
Deriving a feature representation e of an entity pairfAnd after the feature representation u of the sum relation packet, splicing the sum relation packet and the feature representation u to form a new feature vector k ═ ef;u]Feeding into a hidden layer;
the hidden layer is used for receiving the new feature vector, obtaining a final feature representation z of the entity through linear and nonlinear changes of the hidden layer,
z=tanh(Whk+bei) (8)
wherein WhIs a parameter matrix, beiIs an offset, k is from efAnd u are spliced to form a new feature vector.
The output layer is configured to output a final classification result, and specifically includes:
and (3) carrying out linear transformation on the feature vector z obtained by the hidden layer, and obtaining the probability score of each relation category by using SoftMax transformation, wherein the formula (9) is as follows:
o=softmax(Woz+bo) (9)
wherein WoIs a parameter matrix, boIs the offset and o is the output result of the entire network.
Preferably, when the model training is performed, a new entity relationship is obtained from the test corpus to form a triple, and the triple is updated to the knowledge base.
The knowledge graph oriented to the electric power field is constructed, and because the structured data of the electric power text is less, the relationships among the entities need to be extracted from a large amount of unstructured data to construct the knowledge graph. And the problem to be solved is the noise reduction problem of the training set constructed by remote supervised learning. In addition, in the aspect of named entity recognition, the accuracy of entity recognition is improved by adding a professional dictionary in the power field. In conclusion, the method and the device solve the problem of noise reduction in the construction training set of the remote supervised learning, improve the accuracy of entity identification, construct an accurate knowledge map in the power field, construct an accurate knowledge base for the power field, provide a foundation for knowledge discovery in the power field, and lay a foundation for extension of the knowledge map in the power field.
There has been described herein only the preferred embodiments of the invention, but it is not intended to limit the scope, applicability or configuration of the invention in any way. Rather, the detailed description of the embodiments is presented to enable any person skilled in the art to make and use the embodiments. It will be understood that various changes and modifications in detail may be effected therein without departing from the spirit and scope of the invention as defined by the appended claims.
Claims (10)
1. A power domain knowledge graph-oriented construction method is characterized by at least comprising the following steps:
step 1, acquiring structured data and semi-structured and unstructured data in the power field on a network;
step 2, manually screening the obtained structured data through words to serve as the extracted triples and serve as a knowledge base for remote supervised learning; for semi-unstructured data, a natural language processing tool LTP is used for segmenting words of the sentences;
step 3, named entity recognition is carried out through a deep learning Chinese named entity recognition method;
and 4, extracting entity relations by adopting a remote supervised learning method, and realizing construction of the knowledge graph facing the power field.
2. The method according to claim 1, wherein in the step 1, structured data and semi-structured and unstructured data of the electric power field are acquired on a network, and encyclopedia texts of the electric power field are crawled on a breadth-first principle.
3. The method according to claim 1, wherein the step 3 of performing named entity recognition by a deep learning chinese named entity recognition method specifically comprises:
step 3-1, carrying out distributed expression on the words;
step 3-2, carrying out model training by using a deep learning network under supervision;
and 3-3, labeling each word in the sequence by using the context information.
4. The method according to claim 1, wherein the step 3-2 of supervised model training using a deep learning network specifically comprises:
taking a power field data set as a training corpus, and training a word vector through a Skip-gram mode; the training network adopts a three-layer neural network consisting of an input layer, a hidden layer and an output layer, wherein 100 neurons of the hidden layer are set according to experience; the neural network module adopts BilSTM, regards semantic association of words in sentences as a sequence problem, and stores historical information for learning during network training; and (4) considering context correlation, performing sequence annotation by adopting a CRF (domain gradient random access) model, and separating the correlation of an output level.
5. The method according to claim 2, wherein the step 4 of performing entity relationship extraction by using a remote supervised learning method specifically comprises:
step 4-1, entity alignment is carried out, and a relation instance set for training and testing is constructed in an entity alignment mode;
mapping the triple relations in the knowledge base to a training document for entity alignment, and generating a relation instance set Q:
Q={qn|qn=(sm,ei,rk,ej),sm∈D} (1)
wherein e isi、ejIs two entities, rkAs a relationship of two entities in a knowledge base, smFor sentences of entity pairs in corpus D, qnIs the generated relationship instance;
and 4-2, performing intra-sentence relation extraction by adopting a relation extraction model based on an attention mechanism.
6. The method according to claim 5, wherein the step 4-1, performing entity alignment, and constructing the relationship instance set for training and testing in an entity alignment manner, specifically comprises:
step 4-1-1, a mapping step, namely mapping each entity into a sentence of a text, wherein the co-occurrence of a pair of entities in each sentence is used as a relation example, and a plurality of relation examples with the same relation form a relation system;
4-1-2, training, namely aligning entities by using triples extracted from the structured data and encyclopedic texts;
and 4-1-3, a testing step, namely generating candidate relation pairs in pairs by using all entities in the test set in a permutation and combination mode, and generating corresponding test examples and relation examples by using the candidate relation pairs and the test corpus in an entity alignment mode.
7. The method according to claim 5, wherein the step 4-2, using a relation extraction model based on attention mechanism to perform intra-sentence relation extraction,
the attention-based relationship extraction model mainly comprises two parts: a sentence vector representation module and a sentence-level attention mechanism module;
the sentence vector representation module is used for obtaining the characteristic representation of each relation instance in the relation system;
and the sentence-level attention mechanism module is used for measuring the importance degree of each relation instance relative to the relation system.
8. The method of claim 7, wherein in the sentence vector representation module, word vector representation is performed using word2vec method, and word position vector is used to capture relative relationship between words and entities in sentences;
the ith word in the sentenceIs denoted as wiUse ofAndthe expression vocabulary wiWord position vectors from two entities, using tiAs a word wiIs expressed as shown in equation (2):
using BilSTM to obtain the forward state and backward state of each vocabulary, and converting the forward state of each vocabulary into forward stateAnd backward stateThe result of the concatenation is expressed as the state of the vocabulary, as shown in equation (3):
after obtaining the state information of all the words, the vector s of the sentenceiThe representation can be determined jointly by all the lexical states inside it:
9. the method of claim 7, wherein the sentence-level attention mechanism module comprises: the system comprises an attention mechanism calculation unit, an entity feature representation layer, a relation package feature representation layer, a hidden layer and an output layer;
the attention mechanism calculating unit is used for calculating the attention mechanismWeights for different instances in the relationship system are calculated to obtain a vector representation for each relationship system. In the weight calculation process, three kinds of characteristic information of a concept vector, a sentence mark vector and a target relation vector are fused on the basis of a sentence vector, wherein the concept vector eiIncluding descriptor vector ciAnd a superior-inferior token vector qi;
The characteristic representation layer of the relationship system is used for obtaining the characteristics of the relationship system, and the characteristics are jointly determined by the relationship examples in the package; the relation system S is composed of n relation examples, S ═ S1,s2,...,snAnd f, the feature vector u of the relationship packet S can be obtained by formula (5):
wherein, αkIs the weight of the kth relationship instance, skIs the feature vector of the kth relationship instance;
the entity feature representation layer is used for obtaining the abstract features of the entity by using the BilSTM, and specifically comprises the following steps:
the respective probability vectors e of the entities in the relation1And e2Taken together, BilSTM is used to obtain the forward state of each entityAnd the state of the consequentThe forward state and the back term state of the entity are merged as shown in equation (6).
After obtaining the respective state vectors of the entities in the relationship, summing the state vectors as the final feature representation e of the entity pairf;
Deriving a feature representation e of an entity pairfAnd after the feature representation u of the sum relation packet, splicing the sum relation packet and the feature representation u to form a new feature vector k ═ ef;u]Feeding into a hidden layer;
the hidden layer is used for receiving the new feature vector, obtaining a final feature representation z of the entity through linear and nonlinear changes of the hidden layer,
wherein WhIs a parameter matrix, beiIs an offset, k is from efAnd (e) forming a new characteristic vector k ═ e by splicing with uf;u]。
The output layer is configured to output a final classification result, and specifically includes:
and (3) carrying out linear transformation on the feature vector z obtained by the hidden layer, and obtaining the probability score of each relation category by using SoftMax transformation, wherein the formula (9) is as follows:
o=softmax(Woz+bo) (9)
wherein WoIs a parameter matrix, boIs the offset and o is the output result of the entire network.
10. The method according to claim 4, wherein, during the model training, new entity relationship forming triples are obtained from the test corpus and updated to the knowledge base.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201910917049.6A CN110866121A (en) | 2019-09-26 | 2019-09-26 | Knowledge graph construction method for power field |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201910917049.6A CN110866121A (en) | 2019-09-26 | 2019-09-26 | Knowledge graph construction method for power field |
Publications (1)
Publication Number | Publication Date |
---|---|
CN110866121A true CN110866121A (en) | 2020-03-06 |
Family
ID=69652230
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN201910917049.6A Pending CN110866121A (en) | 2019-09-26 | 2019-09-26 | Knowledge graph construction method for power field |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN110866121A (en) |
Cited By (10)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN111552817A (en) * | 2020-04-14 | 2020-08-18 | 国网内蒙古东部电力有限公司 | Electric power scientific and technological achievement knowledge map completion method |
CN111597804A (en) * | 2020-05-15 | 2020-08-28 | 腾讯科技(深圳)有限公司 | Entity recognition model training method and related device |
CN111737496A (en) * | 2020-06-29 | 2020-10-02 | 东北电力大学 | Power equipment fault knowledge map construction method |
CN111931483A (en) * | 2020-06-22 | 2020-11-13 | 中国电力科学研究院有限公司 | Extraction method and device for structuring electric power equipment information |
CN112307767A (en) * | 2020-11-09 | 2021-02-02 | 国网福建省电力有限公司 | Bi-LSTM technology-based regulation and control knowledge modeling method |
CN112364654A (en) * | 2020-11-11 | 2021-02-12 | 安徽工业大学 | Education-field-oriented entity and relation combined extraction method |
CN112860908A (en) * | 2021-01-27 | 2021-05-28 | 云南电网有限责任公司电力科学研究院 | Knowledge graph automatic construction method based on multi-source heterogeneous power equipment data |
CN113255917A (en) * | 2021-07-14 | 2021-08-13 | 国网浙江省电力有限公司杭州供电公司 | Data access and integration method based on electric brain |
CN113378574A (en) * | 2021-06-30 | 2021-09-10 | 武汉大学 | Named entity identification method based on KGANN |
CN115510245A (en) * | 2022-10-14 | 2022-12-23 | 北京理工大学 | Unstructured data oriented domain knowledge extraction method |
Citations (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN110019839A (en) * | 2018-01-03 | 2019-07-16 | 中国科学院计算技术研究所 | Medical knowledge map construction method and system based on neural network and remote supervisory |
CN110110335A (en) * | 2019-05-09 | 2019-08-09 | 南京大学 | A kind of name entity recognition method based on Overlay model |
-
2019
- 2019-09-26 CN CN201910917049.6A patent/CN110866121A/en active Pending
Patent Citations (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN110019839A (en) * | 2018-01-03 | 2019-07-16 | 中国科学院计算技术研究所 | Medical knowledge map construction method and system based on neural network and remote supervisory |
CN110110335A (en) * | 2019-05-09 | 2019-08-09 | 南京大学 | A kind of name entity recognition method based on Overlay model |
Non-Patent Citations (2)
Title |
---|
顾溢: "基于BiLSTM-CRF的复杂中文命名实体识别研究", 中国优秀硕士学位论文全文数据库(电子期刊)信息科技辑, no. 2019, 15 July 2019 (2019-07-15), pages 138 - 1489 * |
顾静航: "面向生物医学领域的实体关系抽取研究", 中国博士学位论文全文数据库(电子期刊)信息科技辑, no. 2018, 15 April 2018 (2018-04-15), pages 138 - 15 * |
Cited By (13)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN111552817A (en) * | 2020-04-14 | 2020-08-18 | 国网内蒙古东部电力有限公司 | Electric power scientific and technological achievement knowledge map completion method |
CN111597804B (en) * | 2020-05-15 | 2023-03-10 | 腾讯科技(深圳)有限公司 | Method and related device for training entity recognition model |
CN111597804A (en) * | 2020-05-15 | 2020-08-28 | 腾讯科技(深圳)有限公司 | Entity recognition model training method and related device |
CN111931483A (en) * | 2020-06-22 | 2020-11-13 | 中国电力科学研究院有限公司 | Extraction method and device for structuring electric power equipment information |
CN111737496A (en) * | 2020-06-29 | 2020-10-02 | 东北电力大学 | Power equipment fault knowledge map construction method |
CN112307767A (en) * | 2020-11-09 | 2021-02-02 | 国网福建省电力有限公司 | Bi-LSTM technology-based regulation and control knowledge modeling method |
CN112364654A (en) * | 2020-11-11 | 2021-02-12 | 安徽工业大学 | Education-field-oriented entity and relation combined extraction method |
CN112860908A (en) * | 2021-01-27 | 2021-05-28 | 云南电网有限责任公司电力科学研究院 | Knowledge graph automatic construction method based on multi-source heterogeneous power equipment data |
CN113378574A (en) * | 2021-06-30 | 2021-09-10 | 武汉大学 | Named entity identification method based on KGANN |
CN113378574B (en) * | 2021-06-30 | 2023-10-24 | 武汉大学 | KGANN-based named entity identification method |
CN113255917A (en) * | 2021-07-14 | 2021-08-13 | 国网浙江省电力有限公司杭州供电公司 | Data access and integration method based on electric brain |
CN115510245A (en) * | 2022-10-14 | 2022-12-23 | 北京理工大学 | Unstructured data oriented domain knowledge extraction method |
CN115510245B (en) * | 2022-10-14 | 2024-05-14 | 北京理工大学 | Unstructured data-oriented domain knowledge extraction method |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN110866121A (en) | Knowledge graph construction method for power field | |
CN110825881A (en) | Method for establishing electric power knowledge graph | |
CN107239446B (en) | A kind of intelligence relationship extracting method based on neural network Yu attention mechanism | |
CN108416065B (en) | Hierarchical neural network-based image-sentence description generation system and method | |
WO2021031480A1 (en) | Text generation method and device | |
CN107315737A (en) | A kind of semantic logic processing method and system | |
CN111026842A (en) | Natural language processing method, natural language processing device and intelligent question-answering system | |
CN107526799A (en) | A kind of knowledge mapping construction method based on deep learning | |
CN105843801B (en) | The structure system of more translation Parallel Corpus | |
CN110321418A (en) | A kind of field based on deep learning, intention assessment and slot fill method | |
CN110489554B (en) | Attribute-level emotion classification method based on location-aware mutual attention network model | |
CN105868187B (en) | The construction method of more translation Parallel Corpus | |
CN111858940A (en) | Multi-head attention-based legal case similarity calculation method and system | |
CN114818717B (en) | Chinese named entity recognition method and system integrating vocabulary and syntax information | |
CN108874896A (en) | A kind of humorous recognition methods based on neural network and humorous feature | |
CN107688583A (en) | The method and apparatus for creating the training data for natural language processing device | |
CN112349294B (en) | Voice processing method and device, computer readable medium and electronic equipment | |
CN116628186B (en) | Text abstract generation method and system | |
CN111523328B (en) | Intelligent customer service semantic processing method | |
CN111597349B (en) | Rail transit standard entity relation automatic completion method based on artificial intelligence | |
CN113656564A (en) | Power grid service dialogue data emotion detection method based on graph neural network | |
CN113901208A (en) | Method for analyzing emotion tendentiousness of intermediate-crossing language comments blended with theme characteristics | |
CN116258147A (en) | Multimode comment emotion analysis method and system based on heterogram convolution | |
Atef et al. | AQAD: 17,000+ arabic questions for machine comprehension of text | |
CN111563148A (en) | Dialog generation method based on phrase diversity |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination |