CN108228728A - A kind of paper network node of parametrization represents learning method - Google Patents
A kind of paper network node of parametrization represents learning method Download PDFInfo
- Publication number
- CN108228728A CN108228728A CN201711308050.6A CN201711308050A CN108228728A CN 108228728 A CN108228728 A CN 108228728A CN 201711308050 A CN201711308050 A CN 201711308050A CN 108228728 A CN108228728 A CN 108228728A
- Authority
- CN
- China
- Prior art keywords
- paper
- node
- queue
- neighbours
- triple
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Granted
Links
- 238000000034 method Methods 0.000 title claims abstract description 47
- 230000008447 perception Effects 0.000 claims abstract description 34
- 238000013528 artificial neural network Methods 0.000 claims abstract description 32
- 238000005070 sampling Methods 0.000 claims abstract description 23
- 238000012549 training Methods 0.000 claims abstract description 19
- 238000005295 random walk Methods 0.000 claims abstract description 14
- 230000001537 neural effect Effects 0.000 claims abstract description 7
- 230000014509 gene expression Effects 0.000 claims description 41
- 230000006870 function Effects 0.000 claims description 40
- 230000000694 effects Effects 0.000 claims description 10
- 241000689227 Cora <basidiomycete fungus> Species 0.000 claims description 9
- 238000012545 processing Methods 0.000 claims description 9
- 238000002474 experimental method Methods 0.000 claims description 5
- 235000008694 Humulus lupulus Nutrition 0.000 claims description 4
- 241000287196 Asthenes Species 0.000 claims description 3
- 230000008569 process Effects 0.000 claims description 3
- 238000013508 migration Methods 0.000 claims 1
- 230000005012 migration Effects 0.000 claims 1
- 230000009466 transformation Effects 0.000 abstract description 2
- 238000011156 evaluation Methods 0.000 description 11
- 230000004913 activation Effects 0.000 description 4
- 238000013507 mapping Methods 0.000 description 4
- 239000011159 matrix material Substances 0.000 description 4
- 230000009191 jumping Effects 0.000 description 3
- 206010038743 Restlessness Diseases 0.000 description 2
- 238000004458 analytical method Methods 0.000 description 2
- 238000000205 computational method Methods 0.000 description 2
- 230000007423 decrease Effects 0.000 description 2
- 210000005036 nerve Anatomy 0.000 description 2
- 238000013519 translation Methods 0.000 description 2
- 230000014616 translation Effects 0.000 description 2
- 241000721047 Danaus plexippus Species 0.000 description 1
- 238000013473 artificial intelligence Methods 0.000 description 1
- 238000010009 beating Methods 0.000 description 1
- 210000003850 cellular structure Anatomy 0.000 description 1
- 238000012512 characterization method Methods 0.000 description 1
- 238000002790 cross-validation Methods 0.000 description 1
- 238000013135 deep learning Methods 0.000 description 1
- 238000011161 development Methods 0.000 description 1
- 235000013399 edible fruits Nutrition 0.000 description 1
- 238000011478 gradient descent method Methods 0.000 description 1
- 238000012886 linear function Methods 0.000 description 1
- 230000006855 networking Effects 0.000 description 1
- 238000013179 statistical model Methods 0.000 description 1
- 238000012795 verification Methods 0.000 description 1
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F16/00—Information retrieval; Database structures therefor; File system structures therefor
- G06F16/30—Information retrieval; Database structures therefor; File system structures therefor of unstructured textual data
- G06F16/36—Creation of semantic tools, e.g. ontology or thesauri
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F16/00—Information retrieval; Database structures therefor; File system structures therefor
- G06F16/30—Information retrieval; Database structures therefor; File system structures therefor of unstructured textual data
- G06F16/33—Querying
- G06F16/3331—Query processing
- G06F16/334—Query execution
- G06F16/3347—Query execution using vector based model
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
- G06N3/045—Combinations of networks
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
- G06N3/048—Activation functions
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/08—Learning methods
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06Q—INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES; SYSTEMS OR METHODS SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES, NOT OTHERWISE PROVIDED FOR
- G06Q50/00—Information and communication technology [ICT] specially adapted for implementation of business processes of specific business sectors, e.g. utilities or tourism
- G06Q50/01—Social networking
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Computational Linguistics (AREA)
- General Engineering & Computer Science (AREA)
- Data Mining & Analysis (AREA)
- General Health & Medical Sciences (AREA)
- Health & Medical Sciences (AREA)
- Computing Systems (AREA)
- Mathematical Physics (AREA)
- Molecular Biology (AREA)
- Evolutionary Computation (AREA)
- Artificial Intelligence (AREA)
- Life Sciences & Earth Sciences (AREA)
- Biomedical Technology (AREA)
- Software Systems (AREA)
- Biophysics (AREA)
- Business, Economics & Management (AREA)
- Databases & Information Systems (AREA)
- Economics (AREA)
- Human Resources & Organizations (AREA)
- Marketing (AREA)
- Primary Health Care (AREA)
- Strategic Management (AREA)
- Tourism & Hospitality (AREA)
- General Business, Economics & Management (AREA)
- Information Retrieval, Db Structures And Fs Structures Therefor (AREA)
Abstract
A kind of paper network node the invention discloses parametrization represents learning method, and this method is built an empty paper node queue, the neighbor node of any one paper node and the neighbor node of neighbours are then sampled using random walk mode first;And using the paper node of selection as first element of paper node queue, subsequent foundation redirects probability and obtains the other elements of paper node queue;Traversal completes all paper nodes, Ze You papers node queue set;Then using the neural metwork training data of positive and negative method of sampling generation multi-layer perception (MLP);It is finally handled using neural network paper probabilistic model, obtains the nonlinear transformation that paper node semantics information is represented to paper knot vector, and then the vector for obtaining paper node represents.
Description
Technical field
The present invention relates to the expression learning methods of paper network, more particularly, refer to a kind of paper network of parametrization
Node table dendrography learning method.
Background technology
Social networks belongs to internet concept noun.Such as the social networking service of Blog, WIKI, Tag, SNS, RSS etc.
Platform.Internet causes a kind of completely new human society tissue and existential mode quietly to come into the present invention, constructs one and surmounts
On terrestrial space, huge group-network colony, the human society of 21 century gradually emerge brand-new form
With speciality, the individual of network global age is being polymerized to new social groups.Paper network refers between paper and paper
Relational network, mutual reference and shared author between paper etc. are shown as on the net.
The expression study of paper network is mostly using non-parametric model, such as " DeepWalk at present:Online
Learning of Social Representations " translations are:Depth is walked:The on-line study that social activity represents, Bryan
Perozzi etc., 26Mar 2014;Using word2vec imparametrizations method to the expression of paper network in the document
It practises.
Network structure refers to the connectivity of network physically.There are many kinds of the topological structures of network, as two-dimensional structure has
Annular, star, tree-like, neighbour connect net and beating pipeline array etc., reference《Interconnection network architecture is analyzed》, Wang Dingxing is old
State is good to write, October nineteen ninety, the 36-38 pages content.With the development of network, also occur such as reticular structure, cellular knot
Structure etc..
The expression learning method of current paper network has to the paper in all paper networks of traversal completion, Cai Nengxue
Practise the expression of paper.The expression study that will cannot carry out newly-increased paper when paper network, which has, newly increases paper, can not be further
The newly-increased paper of completion classification, analysis work.
Invention content
In order to solve the problems, such as that newly-increased paper can not be indicated study, the present invention proposes a kind of paper net of parametrization
Network node table dendrography learning method.In the present invention, star paper network structure is carried out by random walk statistical model first
Sampling, obtains paper knot vector information;The paper node queue that sampling is completed is made of a series of paper node, each
The secondary selection to next paper node is all random;.After paper network samples step is carried out, the present invention constructs one
A deep neural network based on twin network frame, wherein two identical sub-networks of twin network are by multilayer sense
Know machine (MLP) composition, the present invention is using the multi-layer perception (MLP) learnt as nonlinear mapping function, by building from network section
Point rich text information represents that the nonlinear mapping function of vector obtains network node and represents vector to network node.
The paper network node that the present invention proposes a kind of parametrization represents learning method, it is characterised in that includes following
Step:
Step 1, based on random walk method sample obtain any one paper node neighbours-paper set of node and
The neighbours of neighbours-paper set of node;
Step 101:A paper node empty queue is built, is denoted as V, the V is used for storing paper sequence node;Paper section
The maximum queue element digit of point empty queue V is mv, and the value of mv is 10~20;Then step 102 is performed;
Step 102:Choose any one paper node papera, then by the paperaIt is put into paper node queue V
1st;Then step 103 is performed;
Step 103:Acquisition belongs to any one paper node paperaWhole neighbours' paper sets of node, be denoted asNeighbours' paper node refers to and any one paper
Node paperaBetween exist even side paper node set;Then step 104 is performed;
Step 104:According to neighbours' paper set of node
Middle neighbor node sum B determines that jumping to first redirects probabilityC represents hop count;Then step 105 is performed;
Step 105:Using alias sampling algorithm (alias sampling), according to currentDescribedThe middle neighbours' paper node for obtaining next-hopSimultaneously willIt is put into the 2nd of paper node queue V;Then step 106 is performed;
Step 106:Acquisition belongs to neighbours' paper nodeWhole neighbours' paper sets of node, i.e., the neighbours of neighbours-
Paper set of nodeThen step 107 is performed;
Step 107:Calculate neighbours' paper nodeWith any one paper node paperaBetween most short hop countThen step 108 is performed;
WhereinWhat is represented is the fewest number of hops from any one neighbours' paper node to previous paper node
Distance;
Step 108:According to describedTo determineIt jumps to second and redirects probabilityThen
Perform step 109;
Described second redirects probabilityC represents hop count;
Step 109:ThroughAfter determining, according toIt is sampled with alias, selectionAs next-hop
Paper node simultaneously willThe 3rd be put into paper node queue V;Then step 110 is performed;
Step 110:Cycle performs step 106 and step 109, until when the digit in paper node queue V is mv, this
Random walk stops;Then step 111 is performed;
Step 111:Step 101 is repeated to step 110 for each paper node in entire paper network, is come
The neighbor node sampling of paper node is completed, Ze You papers node queue set is denoted as VF={ V1,V2,...,Vf,...,VF};
Then step 201 is performed;
V1Represent first paper node queue;
V2Represent second paper node queue;
VfRepresent any one paper node queue, f represents the identification number of paper node queue;
VFRepresent the last one paper node queue, F represents the sum of paper node queue set, f ∈ F;
Step 2, using the neural metwork training data of negative method of sampling generation multi-layer perception (MLP);
Step 201:Establish positive sample queue QpWith negative sample queue Qn, it is required just that training neural network is stored respectively
Then sampled data and negative sampled data perform step 202;
Step 202:Neighbours window size hyper parameter WD is set up, if WD is in paper node queue VfIn, then belong to paper section
Point queue VfIn each paper be denoted asThen step is performed
203;
Expression belongs to any one paper node queue VfFirst paper node;
Expression belongs to any one paper node queue VfSecond paper node;
Expression belongs to any one paper node queue VfAny one paper node, g represent neighbours' paper
The identification number of node;
Expression belongs to any one paper node queue VfThe last one paper node, G represent paper node
Queue VfLength, g ∈ G;
It represents in adjacent paper nodeThe node of middle minimum identification number;
It represents in adjacent paper nodeThe node of middle maximum identification number;
It represents in adjacent paper nodeIn removeWithIn addition
Queue-adjacent paper node, subscript l expressions are not maximum nor the identification number of minimum paper node;
Step 203:For any one arbitrary queue-paper nodeAccording to the sequence of its neighbours' identification number, from
It is small to being sampled greatly, sampling process is to described
In each node and arbitrary queue-paper nodeA triple is formed, then performs step 204;
For describedWith arbitrary queue-paper nodeForm a triple, i.e., (δ), wherein δ=+ 1 represents the triple as positive sample, otherwise δ=- 1 item represents the ternary
Group is negative sample, and will (δ) it is inserted into positive sample queue QpIn;
For describedWith arbitrary queue-paper nodeForm a triple, i.e., ( δ), wherein δ=+ 1 represents the triple as positive sample, otherwise δ=- 1 item represents the triple as negative sample, and will
(δ) it is inserted into positive sample queue QpIn;
For describedWith arbitrary queue-paper nodeForm a triple, i.e., (δ), wherein δ=+ 1 represents the triple as positive sample, otherwise δ=- 1 item represents the ternary
Group is negative sample, and will (δ) it is inserted into positive sample queue QpIn;
Step 204:Cycle performs step 202 and step 203, until paper node queue set VF={ V1,V2,...,
Vf,...,VFIn all paper node queues in all paper nodes all complete the sampling operation of neighbours' paper node, obtain
To positive sample queue Qp, then perform step 207;
Step 205:Paper nodes all in network are sampled, choose any two paper section from network every time
Point, i.e., the first arbitrary paper node papera, the second arbitrary paper node papero;Connect if existed between two paper nodes
Side or two paper nodes randomly selected are identical, then continue this step, otherwise by any two paper node papera、
paperoForm triple (papera,papero, -1) and deposit negative sample queue QnIn, then perform step 206;
Step 206:Cycle performs step 205, sets up a positive and negative sample proportion parameter μ, it is assumed that positive sample queue QpIn
Triple number is np, then works as QnIn triple quantity be equal to μ × np when stop, then perform step 207;
Step 207:The positive sample queue Q that will be obtained in step 204pWith the negative sample queue Q obtained in step 206nMerge
Together, a new sample queue Q is obtainedNewly={ Q1....,Q(1+μ)×np, it is rear to perform step 208;
Q1Represent new sample queue QNewlyIn minimum identification number triple;
Q(1+μ)×npRepresent new sample queue QNewlyIn maximum identification number triple;Subscript (1+ μ) × np representative samples
Queue QNewlyIn include (1+ μ) × np triple;
Step 208:By new sample queue QNewly={ Q1....,Q(1+μ)×npIn all elements upset sequence, obtain unrest
The sample queue Q of sequenceSequence={ Q1....,Q(1+μ)×np, then perform step 301;
Step 3, the processing in the neural network paper probabilistic model based on multi-layer perception (MLP);
Step 301:The Q obtained for step 208Sequence={ Q1....,Q(1+μ)×np, a triple is selected every time
(papera,papero, δ), it is put into neural network paper probabilistic model and is learnt as a pair of of paper node, perform step
302;
Step 302:For two paper node paper in each tripleaWith papero, using modelIt is mapped, obtains the vector after two corresponding transformationPerform step 303;
To belong to paperaMulti-layer perception (MLP) function;
To belong to paperoMulti-layer perception (MLP) function;
Step 303:The Euclidean distance of two paper nodes is calculated, performs step 304;
Euclidean distance is:
EposRepresent the Euclidean shortest distance;EnegRepresent Euclidean longest distance;C represents hop count;
Step 304:The merging of positive negative sample is put into the loss function of the Euclidean distance represented about paper distribution using δ
In, and the loss function for being balanced positive negative sample calculates, and obtains whole loss function L, performs step 305;
Step 305:Non-linear transform function f is determined using stochastic gradient descent algorithmθ, complete any two paper section
Point paperaWith paperoExpression study.
Network node expression is that each node in network is described with a vector.In order to handle in social networks
Numerous and jumbled information and neighbor node relationship, the present invention propose a kind of parametrization network node represent learning method.It is this
Network node represents that learning method can learn to a nonlinear mapping function so that can be simply from network node
The vector that content information obtains network node represents.The vector of one node is represented, first can obtain it using random walk
All mid-side nodes, then according to the relationship between twin network struction node and its neighbor node, and then learn to determine Nonlinear Mapping
Function.In order to verify the effect of the present invention, the present invention carries out node point using Cora data sets to whole nodes of paper network
The work of class, in the emulation experiment, the network node that is obtained by the method for the present invention represents vector with identical SVM classifier
In the case of, classification results are significantly better than other methods, can verify that the present invention is carrying out network node table to paper network
Show that aspect is effective.
Description of the drawings
Fig. 1 is the flow chart that the paper network node that the present invention parameterizes represents study.
Fig. 2 is the evaluation result of the Micro-F1 indexs in Cora data sets.
Fig. 3 is the evaluation result of the Macro-F1 indexs in Cora data sets.
Fig. 4 is the evaluation result of the Micro-F1 indexs in Wiki data sets.
Fig. 5 is the evaluation result of the Macro-F1 indexs in Wiki data sets.
Specific embodiment
Below in conjunction with drawings and examples, the present invention is described in further detail.
In the present invention, paper is denoted as paper, and more paper paper form a paper set, are denoted as AP, and AP=
{paper1,paper2,…,papera,…,papero,…,paperA};Any one paper in the paper set AP exists
It is known as a paper node in star paper network structure;
paper1Represent first paper node;
paper2Represent second paper node;
paperaRepresent a-th of paper node, a represents the identification number of paper node;
paperARepresent the last one paper node, A represents paper sum, a ∈ A.
For convenience of explanation, paperaAlso referred to as any one paper node, paperoIt is except paperaExcept another
Arbitrary paper node, hereinafter by paperaReferred to as first arbitrary paper node, paperoReferred to as second arbitrary paper node.
Any one paper node paper will be belonged toaWhole neighbours' paper nodes be denoted asAndAlso referred to as neighbours-paper node
Collection;
Expression belongs to any one paper node paperaFirst neighbor node;
Expression belongs to any one paper node paperaSecond neighbor node;
Expression belongs to any one paper node paperaAny one neighbor node, b represent neighbor node
Identification number;
Expression belongs to any one paper node paperaThe last one neighbor node, B expression belong to
paperaNeighbor node sum, B ∈ A.
Any one neighbor node will be belonged toWhole neighbours' paper nodes be denoted asAndThe also referred to as neighbours of neighbours-paper section
Point set.
Expression belongs to any one neighbours' paper nodeFirst neighbor node;
Expression belongs to any one neighbours' paper nodeSecond neighbor node;
Expression belongs to any one neighbours' paper nodeAny one neighbor node, e expression belong to
Neighbours' paper nodeNeighbor node identification number;
Expression belongs to any one neighbours' paper nodeThe last one neighbor node, E expression belong toThe neighbor node sum neighbor node of neighbours (referred to as sum), E ∈ A.
In the present invention, the star paper network structure used for《Interconnection network architecture is analyzed》Fig. 1 .19 of page 37
(c) structure.Wang Dingxing, Chen Guoliang are write;October nineteen ninety the first edition.
In the present invention, paper node semantics information refers to passing through the word that the topic, abstract, text of paper include
Wordization processing carries out vectorial expression.The wordization processing goes out for the paper node semantics information in any one papers contents
The coding of 0 or 1 binaryzation is carried out whether now, so as to obtain the vector of the characterization of the papers contents corresponding 0 or 1." 0 " represents not
Occur, " 1 " represents occur.It is handled using wordization and handled belonging to paper node all in star paper network structure,
Obtain word number and the associated two-dimensional matrix of paper number of nodes, referred to as paper two values matrix.
Using the neural network paper probabilistic model of paper node semantics information architecture multi-layer perception (MLP)
In the present invention, the structure of paper probabilistic model includes:(A) neural network paper probabilistic model expression formula is set(B) from AP={ paper1,paper2,…,papera,…,papero,…,paperAIn choose appoint
Anticipate two paper node paperaWith papero, and by paperaWith papero In mapped, point
It does not obtain belonging to paperaMulti-layer perception (MLP) functionBelong to paperoMulti-layer perception (MLP) function(C) foundationWithCalculate paperaWith paperoBetween Euclidean distance, and be balanced positive negative sample loss function processing;
(D) using stochastic gradient descent algorithm to belonging to fθWeight parameter WEIGHT and offset parameter BIAS processing, learned
Practise the non-linear transform function f of targetθ, all triples are traversed, are obtained to paper node semantics information based on multilayer sense
Know the neural metwork training of machine.
In the present invention, for paperaThere is its corresponding rich text informationThen using multi-layer perception (MLP) god
It is obtained through network describedNonlinear transformation.Assuming that multi-layer perception (MLP) is H layers a total of, the nerve based on multi-layer perception (MLP)
Network has the weight parameter WEIGHT of each layer and offset parameter BIAS.
In the present invention, weight parameter WEIGHT={ weight1,weight2,...,weighth,...,weightH}。
In the present invention, offset parameter BIAS={ bias1,bias2,...,biash,...,biasH}。
weight1Represent the weight parameter of the first layer network in neural network;
weight2Represent the weight parameter of the second layer network in neural network;
weighthRepresent the weight parameter of any one layer network in neural network, h represents the number of plies identification number of perceptron;
weightHRepresent the weight parameter of last layer network in neural network, H represents total number of plies of perceptron;
bias1Represent the offset parameter of the first layer network in neural network;
bias2Represent the offset parameter of the second layer network in neural network;
biashRepresent the offset parameter of any one layer network in neural network;
biasHRepresent the offset parameter of last layer network in neural network.
The output of multi-layer perception (MLP) first layer is denoted asWhereinGeneration
The output of first layer, f in table multi-layer perception (MLP)1Represent the activation primitive of first layer neural network.
Similarly, multi-layer perception (MLP) second layer output is denoted asWherein
Represent the output of the second layer in multi-layer perception (MLP), f2Represent the activation primitive of second layer neural network.
Any one layer of output of multi-layer perception (MLP) is denoted asForfh
Represent the activation primitive of any one layer of neural network.
Last layer of output of multi-layer perception (MLP) is denoted as
In the present invention, the activation primitive f for any one layer in multi-layer perception (MLP)h, generally all select non-linear letter
Number, such as sigmoid tanh functions etc..Output for last layer of multi-layer perception (MLP)It is multiple non-linear
Function is for inputTransformation, therefore can simply portray forWherein θ represents all parametrizations
The summation of function.By described inIt is final as the neural network paper probabilistic model based on multi-layer perception (MLP)
Output, then have
In the present invention, in order to enable the Euclidean distance between point similar in expression of space is short as far as possible, without phase
As Euclidean distance between point it is long as far as possible.Its citation form is:
EposRepresent the Euclidean shortest distance;EnegRepresent Euclidean longest distance;C represents hop count.
In the present invention, due to triple (papera,papero, δ) in δ represent the triple be positive sample or
The mark of negative sample, wherein positive sample are regarded as needing point similar in space, and negative sample may be considered needs and exist
The point of distance as far as possible in space.Therefore, for this application, the present invention can utilize δ to be put into the merging of positive negative sample
In the loss function of Euclidean distance represented about paper distribution:
M represents QSequenceIn any one triple identification number,Represent the first arbitrary paper in triple m
Node,Represent the second arbitrary paper node in triple m, δ(m)Represent the mark of positive negative sample in triple m.L
What is represented is whole loss function, which should be out of order sample sequence QSequenceThe sum of middle all elements loss function.
In the present invention, due to the ratio of positive negative sample difference, and the similitude between positive negative sample is different.Such as
, can be more like probably due to there is even side between positive sample, and negative sample then differs greatly, the loss that negative sample positive in this way generates
Function will differ, therefore for this application, and the present invention needs one to reconcile parameter γ to balance the loss function of positive negative sample,
Therefore loss function will add in γ, become:
In the present invention, the purpose of training neural network is the value of loss function can be reduced at least, in order to right
Neural network is trained, and determines neural network weight and the value of neural network biasing, and the present invention is calculated using stochastic gradient descent
Method carries out the study of network parameter.
In the present invention, to model be trained for non-linear transform function f is determined by stochastic gradient descent algorithmθ, by
In non-linear transform function fθMainly include weight parameter WEIGHT and offset parameter BIAS.For weight parameter WEIGHT and partially
It is L relative to weight parameter WEIGHT and the local derviation of offset parameter BIAS to put updated value that every subgradient of parameter BIAS declines,
Therefore, in each iteration, weight parameter WEIGHT and offset parameter BIAS is carried out more with learning rate ε according to parameter updated value
Newly:
WEIGHTAfterwards=WEIGHTBefore+ε·ΔWEIGHT
BIASAfterwards=BIASBefore+ε·ΔBIAS
WEIGHTBeforeFor the weight parameter of last layer in perceptron, WEIGHTAfterwardsFor the weight parameter of current layer in perceptron,
Δ WEIGHT is to decline per subgradient with L relative to the local derviation of weight parameter WEIGHT.
BLASBeforeFor the offset parameter of last layer in perceptron, BLASAfterwardsFor the offset parameter of current layer in perceptron, Δ
BLAS is to decline per subgradient with L relative to the local derviation of offset parameter BLAS.
When stochastic gradient descent is used, since training iterations are excessive, it may appear that the phenomenon that over-fitting, because
This, present invention employs early-stop (translation is terminates in advance) methods, do not continue to become hour i.e. in training to loss function L
Deconditioning, come the over-fitting occurred when preventing and training." terminating in advance " is《Deep learning》Section 7.8, page 151 interior
Hold, author is her grace Goodfellow, Joshua Ben Jiao etc., translator Zhao Shenjian, multitude Yu monarch;August 1 day first in 2017
Version.
In the present invention, each layer in perceptron of weight parameter WEIGHT and offset parameter BIAS is preserved, is learnt
The non-linear transform function f of targetθ, so as to complete the neural metwork training based on multi-layer perception (MLP), finally obtain according to study
Target fθTo the paperaIt generates it and represents vector, the i.e. nerve for paper node semantics information architecture multi-layer perception (MLP)
Network paper probabilistic model
A kind of paper network node of parametrization proposed by the present invention represents learning method, comprises steps that:
Step 1, based on random walk method sample obtain any one paper node neighbours-paper set of node and
The neighbours of neighbours-paper set of node;
Step 101:A paper node empty queue is built, is denoted as V, the V is used for storing paper sequence node;Paper section
The maximum queue element digit of point empty queue V is mv, and the value of mv is 10~20;Then step 102 is performed;
Step 102:Choose any one paper node papera, then by the paperaIt is put into paper node queue V
1st;Then step 103 is performed;
Step 103:Acquisition belongs to any one paper node paperaWhole neighbours' paper sets of node, be denoted asIn the present invention, neighbours' paper node refers to appointing
Anticipate a paper node paperaBetween exist even side paper node set;Then step 104 is performed;
Step 104:According to neighbours' paper set of node
Middle neighbor node sum B determines to jump to the probability of each neighbours' paper node(referred to as first redirecting probability),C represents hop count;Then step 105 is performed;
Step 105:Using alias sampling algorithm (alias sampling), probability is redirected according to currentInstitute
It statesThe middle neighbours' paper node for obtaining next-hopSimultaneously willIt is put into the 2nd of paper node queue V;Then step 106 is performed;
Step 106:Acquisition belongs to neighbours' paper nodeWhole neighbours' paper sets of node, i.e., the neighbours of neighbours-
Paper set of nodeThen step 107 is performed;
Step 107:Calculate neighbours' paper nodeWith any one paper node paperaBetween most short hop countThen step 108 is performed;
In the present invention, whereinWhat is represented is from any one neighbours' paper node to previous paper node
Fewest number of hops distance, if for example, neighbours' paper nodeTo paper node paperaIt is minimum to need 1 to jump, thenIf neighbours' paper nodeIt is exactly paper node papera, thenAnd so on.
Step 108:According to describedTo determineJump to the probability of each neighbours' paper node(referred to as second redirecting probability);Then step 109 is performed;
Described second redirects probabilityC represents hop count.
In the present invention, most short hop count refer between two paper nodes needed for least hops.
In the present invention, p is adjusts the not paper node in the paper node queue V in random walk method
Second redirect probabilityThe parameter (referred to as jumping out parameter) of size, q are described in random walk method for adjusting
Second of paper node in paper node queue V redirects probabilityThe parameter (referred to as jumping into parameter) of size, p, q control
Make the probability redirected, if it is desired to which random walk is more redirected locally, then p needs to set larger;Conversely, q needs
It sets larger.
Step 109:ThroughAfter determining, according toIt is sampled with alias, selectionAs next-hop opinion
Literary node simultaneously willThe 3rd be put into paper node queue V;Then step 110 is performed;
Step 110:Cycle performs step 106 and step 109, until when the digit in paper node queue V is mv, this
Random walk stops;Then step 111 is performed;
Step 111:In the present invention, step 101 is repeated for each paper node in entire paper network
To step 110, to complete the sampling of the neighbor node of paper node, Ze You papers node queue set is denoted as VF={ V1,V2,...,
Vf,...,VF};Then step 201 is performed.
V1Represent first paper node queue;
V2Represent second paper node queue;
VfRepresent any one paper node queue, f represents the identification number of paper node queue;
VFRepresent the last one paper node queue, F represents the sum of paper node queue set, f ∈ F.
Step 2, using the neural metwork training data of negative method of sampling generation multi-layer perception (MLP);
In the present invention, the paper node queue set that training data workable for generating neural network is obtained for step 1
VF={ V1,V2,...,Vf,...,VF};It removes except the training data in paper node queue set, the present invention can be by
Negative sampling algorithm generates the data needed for training pattern.
Step 201:Establish positive sample queue QpWith negative sample queue Qn, it is required just that training neural network is stored respectively
Then sampled data and negative sampled data perform step 202;
Step 202:Neighbours window size hyper parameter WD is set up, if WD is in paper node queue VfIn, then belong to paper section
Point queue VfIn each paper be denoted asThen step is performed
203;
Expression belongs to any one paper node queue VfFirst paper node;
Expression belongs to any one paper node queue VfSecond paper node;
Expression belongs to any one paper node queue VfAny one paper node (referred to as arbitrary team
Row-paper node), g represents the identification number of neighbours' paper node;
Expression belongs to any one paper node queue VfThe last one paper node, G represent paper node
Queue VfLength, g ∈ G.
It represents in adjacent paper nodeThe node of middle minimum identification number.
It represents in adjacent paper nodeThe node of middle maximum identification number.
It represents in adjacent paper nodeIn removeWithIn addition
Any one paper node, abbreviation queue-adjacent paper node.Subscript l expressions are not maximum nor minimum paper node
Identification number, i.e. the other identifier number except this 2 paper nodes.
Step 203:For any one arbitrary queue-paper nodeAccording to the sequence of its neighbours' identification number, from
It is small to being sampled greatly, sampling process is to described
In each node and arbitrary queue-paper nodeA triple is formed, then performs step 204;
For describedWith arbitrary queue-paper nodeForm a triple, i.e., (δ), wherein δ=+ 1 represents the triple as positive sample, otherwise δ=- 1 item represents the ternary
Group is negative sample, and will (δ) it is inserted into positive sample queue QpIn.
For describedWith arbitrary queue-paper nodeForm a triple, i.e., ( δ), wherein δ=+ 1 represents the triple as positive sample, otherwise δ=- 1 item represents the triple as negative sample, and will
(δ) it is inserted into positive sample queue QpIn.
For describedWith arbitrary queue-paper nodeForm a triple, i.e., (δ), wherein δ=+ 1 represents the triple as positive sample, otherwise δ=- 1 item represents the ternary
Group is negative sample, and will (δ) it is inserted into positive sample queue QpIn.
Step 204:Cycle performs step 202 and step 203, until paper node queue set VF={ V1,V2,...,
Vf,...,VFIn all paper node queues in all paper nodes all complete the sampling operation of neighbours' paper node, obtain
To positive sample queue Qp, then perform step 207;
Step 205:Paper nodes all in network are sampled, choose any two paper node from network every time
(two paper nodes of selection can be adjacent or non-conterminous), i.e., the first arbitrary paper node papera, the
Two arbitrary paper node papero.If there is even side ((paper between two paper nodesa,papero) ∈ E) or two
Identical (the paper of paper node randomly selecteda=papero), then continue this step, otherwise by any two paper node
papera、paperoForm triple (papera,papero, -1) and deposit negative sample queue QnIn, then perform step 206;
Step 206:Cycle performs step 205, sets up a positive and negative sample proportion parameter μ, it is assumed that positive sample queue QpIn
Triple number is np, then works as QnIn triple quantity be equal to μ × np when stop, then perform step 207;
Step 207:The positive sample queue Q that will be obtained in step 204pWith the negative sample queue Q obtained in step 206nMerge
Together, a new sample queue Q is obtainedNewly={ Q1....,Q(1+μ)×np, it is rear to perform step 208;
Q1Represent new sample queue QNewlyIn minimum identification number triple.
Q(1+μ)×npRepresent new sample queue QNewlyIn maximum identification number triple.Subscript (1+ μ) × np representative samples
Queue QNewlyIn include (1+ μ) × np triple.
Step 208:By new sample queue QNewly={ Q1....,Q(1+μ)×npIn all elements upset sequence, obtain unrest
The sample queue Q of sequenceSequence={ Q1....,Q(1+μ)×np, then perform step 301.
Step 3, the processing in the neural network paper probabilistic model based on multi-layer perception (MLP);
Step 301:The Q obtained for step 208Sequence={ Q1....,Q(1+μ)×np, a triple is selected every time
(papera,papero, δ), it is put into neural network paper probabilistic model and is learnt as a pair of of paper node, perform step
302;
Step 302:For two paper node paper in each tripleaWith papero, using modelIt is mapped, obtains the vector after two corresponding transformationPerform step 303;
To belong to paperaMulti-layer perception (MLP) function;
To belong to paperoMulti-layer perception (MLP) function;
Step 303:The Euclidean distance of two paper nodes is calculated, performs step 304;
In the present invention, the purpose of twin network is in order to enable the Euclidean distance between point similar in expression of space is use up
It is possible short, and the Euclidean distance between dissimilar point is long as far as possible.Its citation form is:
EposRepresent the Euclidean shortest distance;EnegRepresent Euclidean longest distance;C represents hop count.
Step 304:The merging of positive negative sample is put into the loss function of the Euclidean distance represented about paper distribution using δ
In, and the loss function for being balanced positive negative sample calculates, and obtains whole loss function L, performs step 305;
Step 305:Non-linear transform function f is determined using stochastic gradient descent algorithmθ, complete any two paper section
Point paperaWith paperoExpression study.
Embodiment 1
The present embodiment employs Cora papers data set and Pubmed knowledge network data sets are learnt and experimental work.
Cora be a paper data set in total containing 2708 paper nodes, comprising 2708 nodes and 5429 sides,
The paper rich text information vector that the corresponding length of each node is 1433, the rich text information vector are represented by 0/1
Word whether there is.Meanwhile each node is associated with a category attribute, total category attribute value number is 7.
Pubmed be a knowledge network data set in total containing 19717 paper nodes, comprising 19717 nodes and
44338 sides, the paper rich text information vector that the corresponding length of each node is 500, the rich text information vector
It whether there is by 0/1 expression word.Meanwhile each node is associated with a category attribute, total category attribute value number
It is 3.
In order to verify validity, main contrast of the present invention performances of the distinct methods in paper node-classification task:
DeepWalk:It employs common Random Walk Algorithm to sample network, then be obtained with word2vec algorithms
The expression of each node in network.(2014DeepWalk:online learning of social
representations[J].Perozzi B,Alrfou R,Skiena S.KDD:701-710.)
TADW:Random walk in DeepWalk is decomposed, dexterously adds the rich text information of node, is used
The mode of matrix multiple obtains the expression of each node in network.(2015,Network representation
learning with rich text information[C]YangC,Zhao D,Zhao D,et al.International
Conference on Artificial Intelligence.AAAI Press,:2111-2117.)
Node2Vec:It is the upgrade version of DeepWalk, employs second order Random Walk Algorithm and network is sampled, then
Representing for each node in network is obtained with word2vec algorithms.(2016,node2vec:Scalable Feature
Learning for Networks[C]//Grover A,Leskovec J.KDD:855.)
Node Forecasting Methodology is selected to carry out the comparison that vector represents effect the present invention.This experiment uses hybrid verification skill
Art (cross-validation) selects SVM classifier to classify in different classification Forecasting Methodologies.
It is Micro-F1 and Macro-F1 respectively to be weighed present invention employs two evaluation indexes.
The wherein computational methods of Macro-F1 are:
Wherein PmacroAnd RmacroThe quasi- rate of macro difference and macro recall ratio are represented respectively.
The computational methods of Micro-F1 are:
Wherein PmicroAnd RmicroThe quasi- rate of elementary errors and micro- recall ratio are represented respectively.
Cora data sets effect as shown in Figures 2 and 3, the present invention with other methods Cora data sets comparison imitate
Fruit, what Fig. 2 was represented is performance of each method under Micro-F1 evaluation indexes, and what Fig. 3 was represented is that each method is commented in Macro-F1
Performance under valency index.The training data that the horizontal axis of two figures represents grader accounts for the percentage of total data.It can from figure
To find out, the method for the present invention network representation learning methods all more several than other under Micro-F1 and Macro-F1 evaluation indexes
Effect will be got well, particular it can be seen that compared to purely by the network information, without the DeepWalk using network node semantic information
With Node2vec algorithms, inventive algorithm is under Micro-F1 and Macro-F1 evaluation indexes, for shared by each training data
Ratio has more than 5% promotion, can show the present invention after converged network nodal information and network topology structure, obtain
To network node represent that vector will significantly be got well than the network node that is obtained using network topological information merely represents vector.Together
When, comparison TADW this combine network node information and the method for network topological information can be seen that side proposed by the present invention
Method still has 3% promotion in two evaluation indexes.
Wiki data sets effect as shown in Figure 4 and Figure 5, it can be seen from the figure that the present invention in Micro-F1 and
It is all got well than the effect of other several network representation learning methods under Macro-F1 evaluation indexes.Due to the classification of Wiki data sets
Quantity will be far more than Cora data sets, it is found that do not use the DeepWalk and Node2vec of network node semantic information
Algorithm, classifying quality is poor, far below the result analyzed using TADW.This declarative semantics occupies master in the data set
Lead effect.Under Micro-F1 and Macro-F1 evaluation indexes, the experimental result obtained for TADW methods has the method for the present invention
2% promotion can show the present invention after converged network nodal information and network topology structure, obtained network node
Represent vector compared to directly using matrix multiple to network node represent that vector will be got well.The present invention can be illustrated in network section
Point represents preferably to be merged with semantic information in jointed with network information, obtains preferably representing vector.
By the analysis of Fig. 2-Fig. 5, these experiments, which embody the present invention, naturally converged network structure to be believed with semantic
Both breaths represent vector, therefore can verify effectiveness of the invention so as to obtain better network node.
Claims (6)
1. a kind of paper network node of parametrization represents learning method, it is characterised in that includes the following steps:
Step 1 samples neighbours-paper set of node and the neighbours for obtaining any one paper node based on random walk method
Neighbours-paper set of node;
Step 101:A paper node empty queue is built, is denoted as V, the V is used for storing paper sequence node;Paper node is empty
The maximum queue element digit of queue V is mv, and the value of mv is 10~20;Then step 102 is performed;
Step 102:Choose any one paper node papera, then by the paperaIt is put into the 1st in paper node queue V
Position;Then step 103 is performed;
Step 103:Acquisition belongs to any one paper node paperaWhole neighbours' paper sets of node, be denoted as
Neighbours' paper node refers to discuss with any one
Literary node paperaBetween exist even side paper node set;Then step 104 is performed;
Step 104:According to neighbours' paper set of node
Middle neighbor node sum B determines to jump to first
Redirect probabilityC represents hop count;Then step 105 is performed;
Step 105:Using alias sampling algorithm (alias sampling), according to currentDescribedThe middle neighbours' paper node for obtaining next-hop
Simultaneously willIt is put into the 2nd of paper node queue V;Then step 106 is performed;
Step 106:Acquisition belongs to neighbours' paper nodeWhole neighbours' paper sets of node, i.e. neighbours neighbours-paper
Set of nodeThen step 107 is performed;
Step 107:Calculate neighbours' paper nodeWith any one paper node paperaBetween most short hop countThen step 108 is performed;
WhereinWhat is represented is the fewest number of hops distance from any one neighbours' paper node to previous paper node;
Step 108:According to describedTo determineIt jumps to second and redirects probabilityThen it performs
Step 109;
Described second redirects probabilityC represents hop count.
Step 109:ThroughAfter determining, according toIt is sampled with alias, selectionAs next-hop paper section
Point simultaneously willThe 3rd be put into paper node queue V;Then step 110 is performed;
Step 110:Cycle performs step 106 and step 109, until when the digit in paper node queue V is mv, this is random
Migration stops;Then step 111 is performed;
Step 111:Step 101 is repeated to step 109 for each paper node in entire paper network, to complete
The neighbor node sampling of paper node, Ze You papers node queue set are denoted as VF={ V1,V2,...,Vf,...,VF};Then
Perform step 201.
V1Represent first paper node queue;
V2Represent second paper node queue;
VfRepresent any one paper node queue, f represents the identification number of paper node queue;
VFRepresent the last one paper node queue, F represents the sum of paper node queue set, f ∈ F.
Step 2, using the neural metwork training data of negative method of sampling generation multi-layer perception (MLP);
Step 201:Establish positive sample queue QpWith negative sample queue Qn, the required positive sampling of training neural network is stored respectively
Then data and negative sampled data perform step 202;
Step 202:Neighbours window size hyper parameter WD is set up, if WD is in paper node queue VfIn, then belong to paper node queue
VfIn each paper be denoted asThen step 203 is performed;
Expression belongs to any one paper node queue VfFirst paper node;
Expression belongs to any one paper node queue VfSecond paper node;
Expression belongs to any one paper node queue VfAny one paper node, g represent neighbours' paper node
Identification number;
Expression belongs to any one paper node queue VfThe last one paper node, G represents paper node queue Vf
Length, g ∈ G.
For the node in any one paper queue in the present invention, it is believed that small with nodal distance in queue
In WD whole nodes be positive sample node.Every time, the 2 × WD belonged to adjacent paper sets of node are first obtained for any one paper node present invention, be denoted as
It represents in adjacent paper nodeThe node of middle minimum identification number.
It represents in adjacent paper nodeThe node of middle maximum identification number.
It represents in adjacent paper nodeIn removeWithTeam in addition
Row-adjacent paper node, subscript l expressions are not maximum nor the identification number of minimum paper node;
Step 203:For any one arbitrary queue-paper nodeAccording to the sequence of its neighbours' identification number, from it is small to
It is sampled greatly, sampling process is to describedIn
Each node and arbitrary queue-paper nodeA triple is formed, then performs step 204;
For describedWith arbitrary queue-paper nodeA triple is formed, i.e.,Wherein δ=+ 1 represents the triple as positive sample, otherwise δ=- 1 item represents to be somebody's turn to do
Triple is negative sample, and willIt is inserted into positive sample queue QpIn.
For describedWith arbitrary queue-paper nodeA triple is formed, i.e., Wherein δ=+ 1 represents the triple as positive sample, otherwise δ=- 1 item represents the triple as negative sample, and
It willIt is inserted into positive sample queue QpIn.
For describedWith arbitrary queue-paper nodeA triple is formed, i.e.,Wherein δ=+ 1 represents the triple as positive sample, on the contrary δ=- 1 item represent this three
Tuple is negative sample, and willIt is inserted into positive sample queue QpIn.
Step 204:Cycle performs step 202 and step 203, until paper node queue set VF={ V1,V2,...,
Vf,...,VFIn all paper node queues in all paper nodes all complete the sampling operation of neighbours' paper node, obtain
To positive sample queue Qp, then perform step 207;
Step 205:Paper nodes all in network are sampled, choose any two paper node from network every time, i.e.,
First arbitrary paper node papera, the second arbitrary paper node papero.If between two paper nodes exist even side or
The paper node that person two randomly selects is identical, then continues this step, otherwise by any two paper node papera、papero
Form triple (papera,papero, -1) and deposit negative sample queue QnIn, then perform step 206;
Step 206:Cycle performs step 205, sets up a positive and negative sample proportion parameter μ, it is assumed that positive sample queue QpMiddle triple
Number is np, then works as QnIn triple quantity be equal to μ × np when stop, then perform step 207;
Step 207:The positive sample queue Q that will be obtained in step 204pWith the negative sample queue Q obtained in step 206nMerge one
It rises, obtains a new sample queue QNewly={ Q1....,Q(1+μ)×np, it is rear to perform step 208;
Q1Represent new sample queue QNewlyIn minimum identification number triple.
Q(1+μ)×npRepresent new sample queue QNewlyIn maximum identification number triple.Subscript (1+ μ) × np representative sample queues
QNewlyIn include (1+ μ) × np triple.
Step 208:By new sample queue QNewly={ Q1....,Q(1+μ)×npIn all elements upset sequence, obtain out of order
Sample queue QSequence={ Q1....,Q(1+μ)×np, then perform step 301.
Step 3, the processing in the neural network paper probabilistic model based on multi-layer perception (MLP);
Step 301:The Q obtained for step 208Sequence={ Q1....,Q(1+μ)×np, a triple is selected every time
(papera,papero, δ), it is put into neural network paper probabilistic model and is learnt as a pair of of paper node, perform step
302;
Step 302:For two paper node paper in each tripleaWith papero, using modelIt is mapped, obtains the vector after two corresponding transformationPerform step 303;
To belong to paperaMulti-layer perception (MLP) function;
To belong to paperoMulti-layer perception (MLP) function;
Step 303:The Euclidean distance of two paper nodes is calculated, performs step 304;
Euclidean distance is:
EposRepresent the Euclidean shortest distance;EnegRepresent Euclidean longest distance;C represents hop count.
Step 304:The merging of positive negative sample is put into the loss function of the Euclidean distance represented about paper distribution using δ,
And the loss function for being balanced positive negative sample calculates, and obtains whole loss function L, performs step 305;
Step 305:Non-linear transform function f is determined using stochastic gradient descent algorithmθ, complete any two paper node
paperaWith paperoExpression study.
2. the paper network node of parametrization according to claim 1 represents learning method, it is characterised in that:Build multilayer
The neural network paper probabilistic model of perceptron includes (A) setting neural network paper probabilistic model expression formula(B) from AP={ paper1,paper2,…,papera,…,papero,…,paperAIn select
Take any two paper node paperaWith papero, and by paperaWith papero In reflected
It penetrates, respectively obtains and belong to paperaMulti-layer perception (MLP) functionBelong to paperoMulti-layer perception (MLP) function
(C) foundationWithCalculate paperaWith paperoBetween Euclidean distance, and be balanced the loss letter of positive negative sample
Number processing;(D) using stochastic gradient descent algorithm to belonging to fθWeight parameter WEIGHT and offset parameter BIAS processing,
Obtain the non-linear transform function f of learning objectiveθ, traverse all triples, obtain to paper node semantics information based on
The neural metwork training of multi-layer perception (MLP).
3. the paper network node of parametrization according to claim 1 represents learning method, it is characterised in that:Step 103,
Step 104 and step 105 realize the acquisition of the 2nd bit element in paper node queue V.
4. the paper network node of parametrization according to claim 1 represents learning method, it is characterised in that:Step 106
The acquisition of element after paper node queue V the 2nd bit elements of relaying is realized to step 110, until reaching mv.
5. the paper network node of parametrization according to claim 1 represents learning method, it is characterised in that:In Cora numbers
3% is promoted according to paper node-classification experiment effect is directed on collection.
6. the paper network node of parametrization according to claim 1 represents learning method, it is characterised in that:In Wiki numbers
2% is promoted according to paper node-classification experiment effect is directed on collection.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201711308050.6A CN108228728B (en) | 2017-12-11 | 2017-12-11 | Parameterized thesis network node representation learning method |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201711308050.6A CN108228728B (en) | 2017-12-11 | 2017-12-11 | Parameterized thesis network node representation learning method |
Publications (2)
Publication Number | Publication Date |
---|---|
CN108228728A true CN108228728A (en) | 2018-06-29 |
CN108228728B CN108228728B (en) | 2020-07-17 |
Family
ID=62653503
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN201711308050.6A Active CN108228728B (en) | 2017-12-11 | 2017-12-11 | Parameterized thesis network node representation learning method |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN108228728B (en) |
Cited By (8)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN109213831A (en) * | 2018-08-14 | 2019-01-15 | 阿里巴巴集团控股有限公司 | Event detecting method and device calculate equipment and storage medium |
CN109376864A (en) * | 2018-09-06 | 2019-02-22 | 电子科技大学 | A kind of knowledge mapping relation inference algorithm based on stacking neural network |
CN109558494A (en) * | 2018-10-29 | 2019-04-02 | 中国科学院计算机网络信息中心 | A kind of scholar's name disambiguation method based on heterogeneous network insertion |
CN111292062A (en) * | 2020-02-10 | 2020-06-16 | 中南大学 | Crowdsourcing garbage worker detection method and system based on network embedding and storage medium |
WO2020248342A1 (en) * | 2019-06-14 | 2020-12-17 | 清华大学 | Hyper-parameter optimization method and apparatus for large-scale network representation learning |
CN112148876A (en) * | 2020-09-23 | 2020-12-29 | 南京大学 | Paper classification and recommendation method |
CN112559734A (en) * | 2019-09-26 | 2021-03-26 | 中国科学技术信息研究所 | Presentation generation method and device, electronic equipment and computer readable storage medium |
CN117648670A (en) * | 2024-01-24 | 2024-03-05 | 润泰救援装备科技河北有限公司 | Rescue data fusion method, electronic equipment, storage medium and rescue fire truck |
Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20130066921A1 (en) * | 2011-09-09 | 2013-03-14 | Sri International | Adaptive ontology |
CN106250438A (en) * | 2016-07-26 | 2016-12-21 | 上海交通大学 | Based on random walk model zero quotes article recommends method and system |
CN106777339A (en) * | 2017-01-13 | 2017-05-31 | 深圳市唯特视科技有限公司 | A kind of method that author is recognized based on heterogeneous network incorporation model |
CN107451596A (en) * | 2016-05-30 | 2017-12-08 | 清华大学 | A kind of classified nodes method and device |
-
2017
- 2017-12-11 CN CN201711308050.6A patent/CN108228728B/en active Active
Patent Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20130066921A1 (en) * | 2011-09-09 | 2013-03-14 | Sri International | Adaptive ontology |
CN107451596A (en) * | 2016-05-30 | 2017-12-08 | 清华大学 | A kind of classified nodes method and device |
CN106250438A (en) * | 2016-07-26 | 2016-12-21 | 上海交通大学 | Based on random walk model zero quotes article recommends method and system |
CN106777339A (en) * | 2017-01-13 | 2017-05-31 | 深圳市唯特视科技有限公司 | A kind of method that author is recognized based on heterogeneous network incorporation model |
Cited By (11)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN109213831A (en) * | 2018-08-14 | 2019-01-15 | 阿里巴巴集团控股有限公司 | Event detecting method and device calculate equipment and storage medium |
CN109376864A (en) * | 2018-09-06 | 2019-02-22 | 电子科技大学 | A kind of knowledge mapping relation inference algorithm based on stacking neural network |
CN109558494A (en) * | 2018-10-29 | 2019-04-02 | 中国科学院计算机网络信息中心 | A kind of scholar's name disambiguation method based on heterogeneous network insertion |
WO2020248342A1 (en) * | 2019-06-14 | 2020-12-17 | 清华大学 | Hyper-parameter optimization method and apparatus for large-scale network representation learning |
CN112559734A (en) * | 2019-09-26 | 2021-03-26 | 中国科学技术信息研究所 | Presentation generation method and device, electronic equipment and computer readable storage medium |
CN112559734B (en) * | 2019-09-26 | 2023-10-17 | 中国科学技术信息研究所 | Brief report generating method, brief report generating device, electronic equipment and computer readable storage medium |
CN111292062A (en) * | 2020-02-10 | 2020-06-16 | 中南大学 | Crowdsourcing garbage worker detection method and system based on network embedding and storage medium |
CN112148876A (en) * | 2020-09-23 | 2020-12-29 | 南京大学 | Paper classification and recommendation method |
CN112148876B (en) * | 2020-09-23 | 2023-10-13 | 南京大学 | Paper classification and recommendation method |
CN117648670A (en) * | 2024-01-24 | 2024-03-05 | 润泰救援装备科技河北有限公司 | Rescue data fusion method, electronic equipment, storage medium and rescue fire truck |
CN117648670B (en) * | 2024-01-24 | 2024-04-12 | 润泰救援装备科技河北有限公司 | Rescue data fusion method, electronic equipment, storage medium and rescue fire truck |
Also Published As
Publication number | Publication date |
---|---|
CN108228728B (en) | 2020-07-17 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN108228728A (en) | A kind of paper network node of parametrization represents learning method | |
Toju et al. | Species-rich networks and eco-evolutionary synthesis at the metacommunity level | |
Kashef et al. | An advanced ACO algorithm for feature subset selection | |
CN112508085B (en) | Social network link prediction method based on perceptual neural network | |
Barman et al. | A Boolean network inference from time-series gene expression data using a genetic algorithm | |
Jokar et al. | Community detection in social networks based on improved label propagation algorithm and balanced link density | |
Cavalcanti et al. | Combining diversity measures for ensemble pruning | |
Li et al. | New dandelion algorithm optimizes extreme learning machine for biomedical classification problems | |
Guendouz et al. | A discrete modified fireworks algorithm for community detection in complex networks | |
Kundu et al. | Fuzzy-rough community in social networks | |
Liang et al. | A novel multiple rule sets data classification algorithm based on ant colony algorithm | |
CN116635866A (en) | Method and system for mining minority class data samples to train a neural network | |
Huang et al. | Identifying node role in social network based on multiple indicators | |
Picard et al. | An application of swarm intelligence to distributed image retrieval | |
CN109308497A (en) | A kind of multidirectional scale dendrography learning method based on multi-tag network | |
Wang et al. | Effective identification of multiple influential spreaders by DegreePunishment | |
Bagavathi et al. | Multi-net: a scalable multiplex network embedding framework | |
Wang et al. | Ppisb: a novel network-based algorithm of predicting protein-protein interactions with mixed membership stochastic blockmodel | |
Bakhshi et al. | Fast evolution of CNN architecture for image classification | |
Nasiri et al. | A node representation learning approach for link prediction in social networks using game theory and K-core decomposition | |
Ma et al. | Discrete and continuous optimization based on hierarchical artificial bee colony optimizer | |
Du et al. | Structural balance in fully signed networks | |
Nadimi-Shahraki et al. | Discrete improved grey wolf optimizer for community detection | |
Lyu et al. | Network intrusion detection based on an efficient neural architecture search | |
Shin et al. | Deep reinforcement learning-based network routing technology for data recovery in exa-scale cloud distributed clustering systems |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
GR01 | Patent grant | ||
GR01 | Patent grant |