CN112948506A - Improved meta-learning relation prediction method based on convolutional neural network - Google Patents

Improved meta-learning relation prediction method based on convolutional neural network Download PDF

Info

Publication number
CN112948506A
CN112948506A CN202110357268.0A CN202110357268A CN112948506A CN 112948506 A CN112948506 A CN 112948506A CN 202110357268 A CN202110357268 A CN 202110357268A CN 112948506 A CN112948506 A CN 112948506A
Authority
CN
China
Prior art keywords
entity
relation
neural network
relationship
data
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN202110357268.0A
Other languages
Chinese (zh)
Inventor
吴涛
朱静
先兴平
许爱东
马红玉
冯柏淋
王宇轩
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Chongqing University of Post and Telecommunications
CSG Electric Power Research Institute
Research Institute of Southern Power Grid Co Ltd
Original Assignee
Chongqing University of Post and Telecommunications
Research Institute of Southern Power Grid Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Chongqing University of Post and Telecommunications, Research Institute of Southern Power Grid Co Ltd filed Critical Chongqing University of Post and Telecommunications
Priority to CN202110357268.0A priority Critical patent/CN112948506A/en
Publication of CN112948506A publication Critical patent/CN112948506A/en
Pending legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/20Information retrieval; Database structures therefor; File system structures therefor of structured data, e.g. relational data
    • G06F16/28Databases characterised by their database models, e.g. relational or object models
    • G06F16/284Relational databases
    • G06F16/285Clustering or classification
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/20Information retrieval; Database structures therefor; File system structures therefor of structured data, e.g. relational data
    • G06F16/28Databases characterised by their database models, e.g. relational or object models
    • G06F16/284Relational databases
    • G06F16/288Entity relationship models
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Databases & Information Systems (AREA)
  • Physics & Mathematics (AREA)
  • Data Mining & Analysis (AREA)
  • General Physics & Mathematics (AREA)
  • General Engineering & Computer Science (AREA)
  • General Health & Medical Sciences (AREA)
  • Computational Linguistics (AREA)
  • Evolutionary Computation (AREA)
  • Biophysics (AREA)
  • Molecular Biology (AREA)
  • Computing Systems (AREA)
  • Biomedical Technology (AREA)
  • Artificial Intelligence (AREA)
  • Mathematical Physics (AREA)
  • Software Systems (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Health & Medical Sciences (AREA)
  • Information Retrieval, Db Structures And Fs Structures Therefor (AREA)

Abstract

The invention belongs to the field of graph data relation prediction, and particularly relates to a relation prediction method for improved meta-learning based on a convolutional neural network, which comprises the following steps: acquiring data to be detected in real time, and converting the data to be detected into ternary group data; inputting the converted triple data into a trained improved meta-learning relationship prediction model to obtain a prediction result of the data to be tested; the method solves the problem that a deep learning model which needs a large amount of data support cannot be used due to less samples, and simultaneously uses the convolutional neural network to further obtain the characteristics of the entity through the neighbors of the entity, thereby improving the calculation efficiency.

Description

Improved meta-learning relation prediction method based on convolutional neural network
Technical Field
The invention belongs to the field of graph data relation prediction, and particularly relates to a relation prediction method for improved meta-learning based on a convolutional neural network.
Background
With the advent of the artificial intelligence era, more and more applications fall on the ground, such as intelligent recommendation, knowledge question answering, assistant decision making and the like, so that a large amount of data resources are generated, implicit relations among a plurality of entities are not discovered, and prediction on possibly existing relations among the entities or discovery and restoration of missing content information are required to be achieved through prediction.
Meanwhile, the deep learning has various complex models at present, characteristics can be well described and fitted in a data set for distribution, but the deep learning is easy to overfit in the face of few sample data sets, and although overfit problems can be avoided by adopting modes such as regularization and data enhancement, the strategies do not provide richer information quantity for the model learning process. However, people can be found not to be restricted by small samples by reflecting the development process of human intelligence and cognition. For example, when a human knows to recognize images of a cat or a dog, the human can quickly extract effective features in the images by only providing a small number of pictures about a certain animal, so that the human has the capability of recognizing the pictures. Similarly, deep learning should accumulate experience from the learning process of tasks with similarity, and take a reverse turn for new similar tasks, so that leann to leann is learning, and the learning method has higher-order generalization capability.
Disclosure of Invention
In order to solve the problems in the prior art, the invention provides a relation prediction method for improved meta-learning based on a convolutional neural network, which comprises the following steps: acquiring data to be detected in real time, and converting the data to be detected into ternary group data; inputting the converted triple data into a trained improved meta-learning relationship prediction model to obtain a prediction result of the data to be tested;
the process of training the improved meta-learning relationship prediction model comprises the following steps:
s1: acquiring original data, converting the original data into triple data, and initializing all the triple data to obtain vector representation of a triple;
s2: dividing the converted triple data to obtain a training set and a test set;
s3: selecting triples in a training set to establish a task area, and defining a support set and a query set according to the task area;
s4: extracting specific relation elements in a support set by adopting a relation element learner;
s5: processing the specific relation element by adopting an embedded learner to obtain a gradient element, and updating the relation element according to the gradient element;
s6: transferring the updated relation elements to a query set, and calculating corresponding score functions; calculating a loss function value of the model according to the corresponding score function;
s7: adjusting the parameters of the model according to the loss function of the model, and finishing the training of the model when the loss function value of the model is minimum;
s8: taking the complete triples in the test set as a support set, and taking the triples to be predicted in the test set as a query set; performing neighbor entity fusion processing on the head entity and the tail entity in the query set to obtain a head entity fusing neighbor node information and a tail entity fusing neighbor node information;
s9: and calculating the score of the query set triple to be predicted according to the head-tail entity pair fused with the neighbor node information, and taking the triple with the highest score as the complementary triple to be predicted.
Preferably, the process of establishing the task area includes: will train set DTrainThe triples with the same relation are added into the same set, and the set is defined as the task T corresponding to the relationr(ii) a Randomly extracting a task T from the task area TrSelecting NSTaking a triple sample as a support set S of the taskrThe rest of NQOne sample as the query set Q of the taskr
Further, the selected one is used as the support set SrSample ratio query set Q ofrThe number of samples in (1) is small; namely: n is a radical ofS<NQ
Preferably, the process of extracting a specific relationship element in the support set by using the relationship element learner includes: embedding and representing a head entity and a tail entity in a support set to obtain a relationship element; inputting the obtained relation elements into an L-layer neural network to extract specific relation information corresponding to each entity; and summing the information of each relationship element, and calculating the average relationship information of each relationship element according to the summed result, wherein the average relationship information is a specific relationship element.
Further, the obtained specific relationship element is:
Figure BDA0003004246940000031
xl=σ(Wlxl-1+bl)
Figure BDA0003004246940000032
preferably, the processing of the specific relationship element by the embedded learner includes: calculating the true values of the triples in the support set and the query set through entity embedding and the relationship elements; calculating a loss function and a gradient element based on the embedded learner according to the true values of the triples; and updating the relation element according to the calculated gradient element.
Preferably, the formula for calculating the score function and the loss function is:
Figure BDA0003004246940000033
Figure BDA0003004246940000034
preferably, the process of obtaining the header entity of the fused neighbor node information and the header entity of the fused neighbor node information: performing neighbor sampling on each head entity and tail entity in the query set to obtain neighbor nodes, and constructing neighbor matrixes according to the neighbor nodes; inputting the neighbor matrix into the embedding layer to obtain a word vector matrix; extracting the characteristics of a word vector matrix by adopting a convolutional neural network; performing pooling treatment on the obtained characteristics to obtain characteristic values; cascading all the pooled feature values to obtain a final feature vector; classifying the final feature vectors to obtain a head entity fusing neighbor node information; the process of obtaining the tail entity fusing the neighbor node information is the same as that of the head entity.
Further, the number of neighbor nodes collected by each node is 20.
The method solves the problem that a deep learning model which needs a large amount of data support cannot be used due to less samples, and simultaneously uses the convolutional neural network to further obtain the characteristics of the entity through the neighbors of the entity, thereby improving the calculation efficiency.
Drawings
FIG. 1 is a flow chart of a prediction method for improved meta-learning based on convolutional neural network in text domain according to the present invention;
FIG. 2 is a process of the present invention for processing data based on the text domain convolutional neural network.
Detailed Description
In order to more clearly illustrate the embodiments of the present invention or the technical solutions in the prior art, the technical solutions in the embodiments of the present invention will be clearly and completely described below with reference to the accompanying drawings, and the described embodiments are only a part of the embodiments of the present invention, but not all of the embodiments. All other embodiments, which can be obtained by a person skilled in the art without inventive effort based on the embodiments of the present invention, are within the scope of the present invention.
A relationship prediction method based on improved meta learning of convolutional neural network, as shown in fig. 1, the method comprising: acquiring data to be detected in real time, and converting the data to be detected into ternary group data; and inputting the converted triple data into a trained improved meta-learning relationship prediction model to obtain a prediction result of the data to be detected.
The process of training the improved meta-learning relationship prediction model comprises the following steps:
s1: acquiring original data, converting the original data into triple data, and initializing all the triple data to obtain vector representation of a triple;
s2: dividing the converted triple data to obtain a training set and a test set;
s3: selecting triples in a training set to establish a task area, and defining a support set and a query set according to the task area;
s4: extracting specific relation elements in a support set by adopting a relation element learner;
s5: processing the specific relation element by adopting an embedded learner to obtain a gradient element, and updating the relation element according to the gradient element;
s6: transferring the updated relation elements to a query set, and calculating corresponding score functions; calculating a loss function value of the model according to the corresponding score function;
s7: adjusting the parameters of the model according to the loss function of the model, and finishing the training of the model when the loss function value of the model is minimum;
s8: taking the complete triples in the test set as a support set, and taking the triples to be predicted in the test set as a query set; performing neighbor entity fusion processing on the head entity and the tail entity in the query set to obtain a head entity fusing neighbor node information and a tail entity fusing neighbor node information;
s9: and calculating the score of the query set triple to be predicted according to the head-tail entity pair fused with the neighbor node information, and taking the triple with the highest score as the complementary triple to be predicted.
The process of establishing the task area comprises the following steps: will train set DTrainThe triples with the same relation are added into the same set, and the set is defined as the task T corresponding to the relationr(ii) a In the task areaRandomly extracting a task T from TrSelecting NSTaking a triple sample as a support set S of the taskrThe rest of NQOne sample as the query set Q of the taskr(ii) a Selected as support set SrSample ratio query set Q ofrThe number of samples in (1) is small; namely: n is a radical ofS<NQ
The process of extracting the specific relationship element in the support set by adopting the relationship element learner comprises the following steps: embedding and representing a head entity and a tail entity in a support set to obtain a relationship element; inputting the obtained relation elements into an L-layer neural network to extract specific relation information corresponding to each entity; and summing the information of each relationship element, and calculating the average relationship information of each relationship element according to the summed result, wherein the average relationship information is a specific relationship element.
The process of processing the specific relationship element by adopting the embedded learner comprises the following steps: calculating the true values of the triples in the support set and the query set through entity embedding and the relationship elements; calculating a loss function and a gradient element based on the embedded learner according to the true values of the triples; and updating the relation element according to the calculated gradient element.
As shown in fig. 2, the process of obtaining the header entity of the merged neighbor node information and the header entity of the merged neighbor node information: performing neighbor sampling on each head entity and tail entity in the query set to obtain neighbor nodes, and constructing neighbor matrixes according to the neighbor nodes; inputting the neighbor matrix into the embedding layer to obtain a word vector matrix; extracting the characteristics of a word vector matrix by adopting a convolutional neural network; performing pooling treatment on the obtained characteristics to obtain characteristic values; cascading all the pooled feature values to obtain a final feature vector; classifying the final feature vectors to obtain a head entity fusing neighbor node information; the process of obtaining the tail entity fusing the neighbor node information is the same as that of the head entity.
Preferably, the number of neighbor nodes collected by each node is 20.
A specific implementation mode of a relation prediction method based on improved meta learning of a convolutional neural network comprises the following processes:
step 1: extracting entity relation triples in a relevant range, and defining the entity relation triples as a training set D of a task to be predictedTrainAnd each training set has a separate support set and query set.
Step 2: defining the triples of samples to be predicted as a test set D of prediction tasksTest. Where the test set and training set are very similar but are a single set.
And step 3: to DTrainAnd DTestAll the triples in the triples are initialized to obtain a triplet (h, r, t) represented by a vector, wherein h is a head entity vector, r is a relation vector, t is a tail entity vector, and (h, r, t) belongs to G, G is a triplet sample set, h, t belongs to E, and E is an entity set.
The data is expressed in a triple form and comprises a head entity, a tail entity and a relation between the head entity and the tail entity; and initializing the triple data to obtain the vector form representation of the triple data. Vector representation of triple data can be more easily identified and operated on.
And 4, step 4: training set DTrainThe triples with the same relation are added into the same set and defined as the task T corresponding to the relationr
And 5: randomly extracting a task T from the task area TrTaking out NSTaking a triple sample as a support set S of the taskrThe rest of NQOne sample as the query set Q of the taskrIn which N isS<NQ
Step 6: to get from the supporting set SrA relation element learner which is used for extracting the relation elements and defines a neural network which can be regarded as simple is used for the support set SrThe matching of the head entity to the relationship element in (1), which inputs the support set SrHead and tail entity pairs. Firstly, extracting entity pair specific relation elements through a full-connection neural network of an L layer:
Figure BDA0003004246940000061
xl=σ(Wlxl-1+bl)
Figure BDA0003004246940000062
wherein h isiRepresenting a d-dimensional vector, tiRepresenting a d-dimensional vector, L is the number of layers of the neural network.
And 7: weighting the relation elements of K, wherein the weighted formula is as follows:
Figure BDA0003004246940000063
wherein K represents the number of the relation elements,
Figure BDA0003004246940000064
represents TrThe relationship element of (1).
And 8: the definition embedding learner is used for obtaining gradient elements to rapidly update the relation elements. At task TrIn (1), calculating a support set S according to an entity score calculation formularScore of each entity pair in:
Figure BDA0003004246940000071
calculating a loss score from the scores of each entity pair:
Figure BDA0003004246940000072
wherein the content of the first and second substances,
Figure BDA0003004246940000073
representing a support set SrIth entity pair score, hiRepresenting the ith head entity vector, tiRepresenting the ith tail entity vector,
Figure BDA0003004246940000074
represents TrA relation element of (1), TrIndicating a random extraction of a task, L (S), from the task area Tr) Represents the loss score, S, of each entity pairrRepresenting a support set and gamma representing a boundary hyperparameter, i.e. a tolerable error boundary. When the loss function value is smaller, the representation effect is better, and the representation effect represents the true value of the triple which can be coded by the current model.
And step 9: the gradient parameter shows how the parameter is updated, so that the relation element obtained based on the loss function formula is used as the gradient element, the partial derivative of the loss function is solved, and the relation element is obtained according to the relation element formula after the gradient is reduced:
Figure BDA0003004246940000075
wherein the content of the first and second substances,
Figure BDA0003004246940000076
represents TrThe gradient elements of (a) are selected,
Figure BDA0003004246940000077
denotes the gradient, L (S)r) Represents the loss function, SrA support set is represented.
And carrying out quick iteration on the relation element, wherein the iterative formula is as follows:
Figure BDA0003004246940000078
wherein, beta represents the number of iteration steps,
Figure BDA0003004246940000079
represents TrThe gradient elements of (a) are selected,
Figure BDA00030042469400000710
represents TrThe relationship element of (1). Repeating the steps 6-9, and continuously updating R' until all the support sets S are traversedr
Step 10: transferring the obtained updated relation elements to a query set Qr, and calculating corresponding score function values:
Figure BDA00030042469400000711
calculating a loss function value according to the score value to update the relation element:
Figure BDA00030042469400000712
step 11: selecting the test set D in step 2TestAnd taking the complete triple in the test set as a support set, selecting the relationship element R finally obtained from the step 5 to the step 10, and initializing the relationship in the triple according to the selected relationship element to obtain the initialized triple vector representation.
Step 12: test set DTestAnd taking the triple to be predicted as an inquiry set, sampling the neighbor of each head entity or tail entity, and constructing a neighbor matrix. To efficiently construct the neighbor matrix such that each node samples approximately 20 neighbor nodes, the neighbor matrix is constructed from the neighbor nodes. If the number of the node neighbors is less than 20, a sampling method with a return is adopted until 20 vertexes are sampled; if the number of node neighbors is more than 20, then non-return sampling is adopted.
Step 13: the Embedding layer is input by a neighbor matrix, and because of less samples, word vectors are pre-trained (pre-train) by using external linguistic data and then input to the Embedding layer, and the Embedding layer is initialized by using the pre-trained word vector matrix.
Step 14: after the embedding layer, a convolutional neural network is used to extract features, as shown in fig. 2. Since the relevance of the neighboring neighbor vectors in the neighbor matrix w is relatively high, a one-dimensional convolution is adopted. Wherein, the width of the convolution kernel is the dimension d of the neighbor vector, and the height H is a hyper-parameter, then the formula for performing convolution operation is:
oi=w·A[i:i+H-1],i=1,2,.......,T-H+1
wherein o isiRepresenting the convolution kernel characteristics, w representing the neighbor matrix, i representing the number of rows of the matrix, H representing the height of the matrix, and T representing the number of sampled neighbor nodes.
Preferably, T is the number of sampled neighbors 20. And (3) superposing the bias b, and activating by using an activation function f to obtain the required characteristics, wherein the formula of activating by using the activation function f is as follows:
ci=f(oi+b)
then for a convolution kernel, the characteristic c epsilon R can be obtainedS-H+1Total T-H +1 features. More convolution kernels with different heights can be used, and richer feature expressions can be obtained.
Step 15: the resulting features (feature maps) for different sizes of convolution kernel are different, so a pooling function of 1-max power is used for each feature map, so that the resulting feature for each convolution kernel is a value. And after the characteristic values of all the convolution kernels are passed, cascading the characteristic values to obtain a final characteristic vector.
Step 16: inputting the final feature vector into a fullcontect layer for classification, simultaneously using drop out to prevent overfitting, obtaining head entity or tail entity embedding of fused neighbor node information, and then obtaining a new test set D'Test. As shown in the diagram x, the neighbor node information is embedded into a head entity or a tail entity of the merged neighbor node information through a convolutional neural network (TextCNN).
And step 17: test set D'TestAnd (4) taking the triple to be predicted as an inquiry set, calculating the score of the triple of the inquiry set according to a score calculation formula by using the vector of the triple obtained in the step (11), and taking the triple with the highest score as a complementary triple to be predicted.
The above-mentioned embodiments, which further illustrate the objects, technical solutions and advantages of the present invention, should be understood that the above-mentioned embodiments are only preferred embodiments of the present invention, and should not be construed as limiting the present invention, and any modifications, equivalents, improvements, etc. made within the spirit and principle of the present invention should be included in the protection scope of the present invention.

Claims (9)

1. A relation prediction method of improved meta-learning based on a convolutional neural network is characterized by comprising the following steps: acquiring data to be detected in real time, and converting the data to be detected into ternary group data; inputting the converted triple data into a trained improved meta-learning relationship prediction model to obtain a prediction result of the data to be tested;
the process of training the improved meta-learning relationship prediction model comprises the following steps:
s1: acquiring original data, converting the original data into triple data, and initializing all the triple data to obtain vector representation of a triple;
s2: dividing the converted triple data to obtain a training set and a test set;
s3: selecting triples in a training set to establish a task area, and defining a support set and a query set according to the task area;
s4: extracting specific relation elements in a support set by adopting a relation element learner;
s5: processing the specific relation element by adopting an embedded learner to obtain a gradient element, and updating the relation element according to the gradient element;
s6: transferring the updated relation elements to a query set, and calculating corresponding score functions; calculating a loss function of the model according to the corresponding score function;
s7: adjusting the parameters of the model according to the loss function of the model, and finishing the training of the model when the loss function value of the model is minimum;
s8: taking the complete triples in the test set as a support set, and taking the triples to be predicted in the test set as a query set; performing neighbor entity fusion processing on the head entity and the tail entity in the query set to obtain a head entity fusing neighbor node information and a tail entity fusing neighbor node information;
s9: and calculating the score of the query set triple to be predicted according to the head-tail entity pair fused with the neighbor node information, and taking the triple with the highest score as the complementary triple to be predicted.
2. The method of claim 1, wherein the process of establishing the task area comprises: will train set DTrainThe triples with the same relation are added into the same set, and the set is defined as the task T corresponding to the relationr(ii) a Randomly extracting a task T from the task area TrSelecting NSTaking a triple sample as a support set S of the taskrThe rest of NQOne sample as the query set Q of the taskr
3. The method of claim 2, wherein the selected support set S is a convolutional neural network-based relationship prediction method for improved meta-learningrSample ratio query set Q ofrThe number of samples in (1) is small; namely: n is a radical ofS<NQ
4. The method of claim 1, wherein the extracting of the specific relationship element in the support set by the relationship element learner comprises: embedding and representing a head entity and a tail entity in a support set to obtain a relationship element; inputting the obtained relation elements into an L-layer neural network to extract specific relation information corresponding to each entity; and summing the information of each relationship element, and calculating the average relationship information of each relationship element according to the summed result, wherein the average relationship information is a specific relationship element.
5. The method of claim 4, wherein the obtained specific relation elements are:
Figure FDA0003004246930000021
xl=σ(Wlxl-1+bl)
Figure FDA0003004246930000022
wherein x is0Represents the initial relationship element, hiRepresents the ith head entity vector and the ith head entity vector,
Figure FDA0003004246930000023
denotes performing an XOR operation, tiDenotes the ith tail entity vector, xlRepresents a specific relationship element of the l-th layer neural network, l represents the l-th layer neural network, sigma (.) represents the LeakyRelu activation function, WlRepresenting the l-th layer weight of the neural network, blRepresents the layer i bias term of the neural network,
Figure FDA0003004246930000024
represents the ith entity to the relationship element, WLRepresents the Lth layer weight of the neural network, L represents the total number of layers of the neural network, xL-1A relationship element representing the L-1 th layer, bLRepresents the L-th layer bias term of the neural network.
6. The method of claim 1, wherein the processing of the specific relation element by the embedded learner comprises: calculating the true values of the triples in the support set and the query set through entity embedding and the relationship elements; calculating a loss function and a gradient element based on the embedded learner according to the true values of the triples; and updating the relation element according to the calculated gradient element.
7. The method of claim 1, wherein the formula for calculating the score function and the loss function is as follows:
Figure FDA0003004246930000031
Figure FDA0003004246930000032
wherein the content of the first and second substances,
Figure FDA0003004246930000033
representing a support set SrIth entity pair score, hiRepresenting the ith head entity vector, tiRepresenting the ith tail entity vector,
Figure FDA0003004246930000034
represents TrA relation element of (1), TrIndicating a random extraction of a task, L (S), from the task area Tr) Represents the loss score, S, of each entity pairrRepresenting a support set and gamma representing a boundary hyperparameter, i.e. a tolerable error boundary.
8. The convolutional neural network-based relationship prediction method for improved meta-learning, as claimed in claim 1, wherein the process of obtaining the header entity of the fused neighbor node information and the header entity of the fused neighbor node information: performing neighbor sampling on each head entity and tail entity in the query set to obtain neighbor nodes, and constructing neighbor matrixes according to the neighbor nodes; inputting the neighbor matrix into the embedding layer to obtain a word vector matrix; extracting the characteristics of a word vector matrix by adopting a convolutional neural network; performing pooling treatment on the obtained characteristics to obtain characteristic values; cascading all the pooled feature values to obtain a final feature vector; classifying the final feature vectors to obtain a head entity fusing neighbor node information; the process of obtaining the tail entity fusing the neighbor node information is the same as that of the head entity.
9. The method of claim 8, wherein the number of neighbor nodes collected by each node is 20.
CN202110357268.0A 2021-04-01 2021-04-01 Improved meta-learning relation prediction method based on convolutional neural network Pending CN112948506A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202110357268.0A CN112948506A (en) 2021-04-01 2021-04-01 Improved meta-learning relation prediction method based on convolutional neural network

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202110357268.0A CN112948506A (en) 2021-04-01 2021-04-01 Improved meta-learning relation prediction method based on convolutional neural network

Publications (1)

Publication Number Publication Date
CN112948506A true CN112948506A (en) 2021-06-11

Family

ID=76232096

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202110357268.0A Pending CN112948506A (en) 2021-04-01 2021-04-01 Improved meta-learning relation prediction method based on convolutional neural network

Country Status (1)

Country Link
CN (1) CN112948506A (en)

Citations (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109753570A (en) * 2019-01-11 2019-05-14 中山大学 A kind of scene map vectorization method based on Horn logical AND figure neural network
CN111046193A (en) * 2019-12-27 2020-04-21 南京邮电大学 Domain knowledge graph completion method based on meta-learning
CN111178543A (en) * 2019-12-30 2020-05-19 广东石油化工学院 Probability domain generalization learning method based on meta-learning
CN111291139A (en) * 2020-03-17 2020-06-16 中国科学院自动化研究所 Attention mechanism-based knowledge graph long-tail relation completion method
CN111639196A (en) * 2020-06-03 2020-09-08 核工业湖州工程勘察院有限公司 Multi-layer gradually-enhanced ground disaster knowledge graph and automatic completion method thereof
CN111931506A (en) * 2020-05-22 2020-11-13 北京理工大学 Entity relationship extraction method based on graph information enhancement
CN111949764A (en) * 2020-08-18 2020-11-17 桂林电子科技大学 Knowledge graph completion method based on bidirectional attention mechanism
CN112528035A (en) * 2020-07-22 2021-03-19 中国人民解放军国防科技大学 Knowledge graph reasoning method and device based on relational attention and computer equipment

Patent Citations (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109753570A (en) * 2019-01-11 2019-05-14 中山大学 A kind of scene map vectorization method based on Horn logical AND figure neural network
CN111046193A (en) * 2019-12-27 2020-04-21 南京邮电大学 Domain knowledge graph completion method based on meta-learning
CN111178543A (en) * 2019-12-30 2020-05-19 广东石油化工学院 Probability domain generalization learning method based on meta-learning
CN111291139A (en) * 2020-03-17 2020-06-16 中国科学院自动化研究所 Attention mechanism-based knowledge graph long-tail relation completion method
CN111931506A (en) * 2020-05-22 2020-11-13 北京理工大学 Entity relationship extraction method based on graph information enhancement
CN111639196A (en) * 2020-06-03 2020-09-08 核工业湖州工程勘察院有限公司 Multi-layer gradually-enhanced ground disaster knowledge graph and automatic completion method thereof
CN112528035A (en) * 2020-07-22 2021-03-19 中国人民解放军国防科技大学 Knowledge graph reasoning method and device based on relational attention and computer equipment
CN111949764A (en) * 2020-08-18 2020-11-17 桂林电子科技大学 Knowledge graph completion method based on bidirectional attention mechanism

Similar Documents

Publication Publication Date Title
Chen et al. Shallowing deep networks: Layer-wise pruning based on feature representations
CN111612066B (en) Remote sensing image classification method based on depth fusion convolutional neural network
US8862527B2 (en) Neural networks and method for training neural networks
CN109063113B (en) Rapid image retrieval method, retrieval model and model construction method based on asymmetric depth discrete hash
CN112766280A (en) Remote sensing image road extraction method based on graph convolution
CN109886401A (en) A kind of complex network representative learning method
CN113190688A (en) Complex network link prediction method and system based on logical reasoning and graph convolution
Jha et al. Extracting low‐dimensional psychological representations from convolutional neural networks
CN111461907A (en) Dynamic network representation learning method oriented to social network platform
CN116402133B (en) Knowledge graph completion method and system based on structure aggregation graph convolutional network
Irfan et al. Brain inspired lifelong learning model based on neural based learning classifier system for underwater data classification
CN114548256A (en) Small sample rare bird identification method based on comparative learning
Sokkhey et al. Development and optimization of deep belief networks applied for academic performance prediction with larger datasets
CN114332075A (en) Rapid structural defect identification and classification method based on lightweight deep learning model
Li et al. Performance analysis of fine-tune transferred deep learning
CN116434347B (en) Skeleton sequence identification method and system based on mask pattern self-encoder
CN109934281B (en) Unsupervised training method of two-class network
CN112948506A (en) Improved meta-learning relation prediction method based on convolutional neural network
Laleh et al. Chaotic continual learning
CN115131605A (en) Structure perception graph comparison learning method based on self-adaptive sub-graph
CN113408652A (en) Semi-supervised learning image classification method based on group representation features
CN113205175A (en) Multi-layer attribute network representation learning method based on mutual information maximization
Botteghi et al. Low-dimensional state and action representation learning with mdp homomorphism metrics
Yamashita et al. SWAP-NODE: A regularization approach for deep convolutional neural networks
CN110210988B (en) Symbolic social network embedding method based on deep hash

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
RJ01 Rejection of invention patent application after publication

Application publication date: 20210611

RJ01 Rejection of invention patent application after publication