CN109117943B - Method for enhancing network representation learning by utilizing multi-attribute information - Google Patents

Method for enhancing network representation learning by utilizing multi-attribute information Download PDF

Info

Publication number
CN109117943B
CN109117943B CN201810820414.7A CN201810820414A CN109117943B CN 109117943 B CN109117943 B CN 109117943B CN 201810820414 A CN201810820414 A CN 201810820414A CN 109117943 B CN109117943 B CN 109117943B
Authority
CN
China
Prior art keywords
network
semi
attribute information
information
depth model
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN201810820414.7A
Other languages
Chinese (zh)
Other versions
CN109117943A (en
Inventor
乔立升
陈恩红
刘淇
徐童
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
University of Science and Technology of China USTC
Original Assignee
University of Science and Technology of China USTC
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by University of Science and Technology of China USTC filed Critical University of Science and Technology of China USTC
Priority to CN201810820414.7A priority Critical patent/CN109117943B/en
Publication of CN109117943A publication Critical patent/CN109117943A/en
Application granted granted Critical
Publication of CN109117943B publication Critical patent/CN109117943B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • G06N3/084Backpropagation, e.g. using gradient descent

Abstract

The invention discloses a method for enhancing network representation by utilizing multi-attribute information, which comprises the following steps: fusing different types of attribute similarity information of the network nodes with structural adjacency information between corresponding network nodes through a first semi-supervised depth model to obtain representations of a plurality of different hidden spaces; optimizing structural adjacency information among the network nodes through a second semi-supervised depth model to obtain the representation of the network nodes in a hidden space, which is obtained only by using the structural adjacency information; and fusing the obtained representations of the hidden space to obtain the final network representation. The method can capture the nonlinear network structure, can solve the problem of non-consistency between the network topology structure and each attribute information to a certain extent from a certain angle, fully exerts the enhancement effect of the attribute information and keeps as much information as possible.

Description

Method for enhancing network representation learning by utilizing multi-attribute information
Technical Field
The invention relates to the technical field of machine learning and network representation optimization, in particular to a method for enhancing network representation learning by utilizing multi-attribute information.
Background
Many realistically important networks, such as social networks, reference networks, and airline networks, have complex structural features and rich attribute information (e.g., gender, reviews, user portrayal, etc.). How to utilize this information to better learn about network characterization is a fundamental but challenging task in network mining. One popular and effective way to represent network nodes is to map them to a low dimensional space, which is suitable for a variety of tasks, such as node classification, community discovery, and targeted advertising.
Recently, many algorithms have been proposed for network characterization, such as deep walk, LINE and node2 vec. While these approaches have proven effective in a number of network analysis tasks, most of them primarily consider structural topology information in network characterization learning. In reality, however, the network nodes may have a number of attributes that may be of different types that may help to improve the learning of the network characterization and benefit the subsequent analysis task. In order to obtain a representation with a better effect, some work considers the introduction of information such as node attributes or community structures to learn network representations.
Although some work has made many important efforts in learning to introduce attribute information for node characterization, how to effectively characterize nodes of a network with sparse connections and non-uniform attribute information is still in an ongoing research phase. Generally, in an actual network, connection information between nodes may be missing, attributes may be heterogeneous with respect to each other, and some newly added nodes may have limited or no connection information. In addition, the common attributes that different communities have may be the same, which means that the node signatures learned from network attribute information of different parts of the network node data, the encoded network structure information is likely to have inconsistencies, and vice versa between the attribute information and the structure topology. How to deal with these inconsistencies when we use attribute information to enhance the learning of network representations. The above problems bring challenges to the existing methods for effectively learning network characterization. There have been some efforts to improve the learning effect of network representation by using texts related to nodes, and some efforts to enhance the learning of network representation by using node attributes in a deep learning manner, but they either fail to address the above-mentioned non-uniformity problem in a targeted manner, or mainly get a common information part between different perspective attribute information.
Therefore, it is highly desirable and necessary to explore more technical approaches that can effectively capture nonlinear structural information and overcome the effects of inconsistencies, and retain as much useful information as possible.
Disclosure of Invention
The invention aims to provide a method for enhancing network representation learning by utilizing multi-attribute information, which can capture a nonlinear network structure, solve the problem of non-consistency between a network topology structure and each attribute information to a certain extent from a certain angle, fully play the role of enhancing the attribute information and keep as much information as possible.
The purpose of the invention is realized by the following technical scheme:
a method for enhancing network characterization learning using multi-attribute information, comprising:
s1, fusing different types of attribute similarity information of the network nodes with structural adjacency information between corresponding network nodes through a first semi-supervised depth model to obtain representations of a plurality of different hidden spaces;
step S2, optimizing the structural adjacency information among the network nodes through a second semi-supervised depth model to obtain the representation of the network nodes in the hidden space, which is obtained only by using the structural adjacency information;
and S3, fusing the representations of the hidden space obtained in the steps S1-S2 to obtain the final network representation.
According to the technical scheme provided by the invention, when the representation vector of the network node is learned, the topological structure information and the node attribute information of the network are considered, and meanwhile, the auxiliary function of the attribute information is fully exerted; the scheme is particularly suitable for learning the representation of the network with sparse connection and inconsistent attribute information, and a better network representation learning effect is obtained in the mode; the scheme can provide more accurate support for tasks such as node classification, connection prediction, community discovery and targeted advertising.
Drawings
In order to more clearly illustrate the technical solutions of the embodiments of the present invention, the drawings needed to be used in the description of the embodiments are briefly introduced below, and it is obvious that the drawings in the following description are only some embodiments of the present invention, and it is obvious for those skilled in the art to obtain other drawings based on the drawings without creative efforts.
FIG. 1 is a flowchart of a method for enhancing learning by using multi-attribute information for network characterization according to an embodiment of the present invention;
fig. 2 is an overall block diagram of a model framework for implementing a method for enhancing network characterization learning by using multi-attribute information according to an embodiment of the present invention.
Detailed Description
The technical solutions in the embodiments of the present invention are clearly and completely described below with reference to the drawings in the embodiments of the present invention, and it is obvious that the described embodiments are only a part of the embodiments of the present invention, and not all embodiments. All other embodiments, which can be derived by a person skilled in the art from the embodiments of the present invention without making any creative effort, shall fall within the protection scope of the present invention.
The embodiment of the invention provides a method for enhancing network representation learning by utilizing multi-attribute information, which is mainly used for solving the problems of sparsity and information inconsistency of a network, designs a semi-supervised depth model for combining network topology structure information and node attribute information, and designs a model framework for fusing multi-view information. The design aims to take the topological structure information and the node attribute information of the network into consideration when the characterization vectors of the network nodes are learned, and simultaneously, the auxiliary function of the attribute information is fully exerted.
FIG. 1 is a flow chart of a method for enhancing network characterization learning using multi-attribute information, and FIG. 2 is a corresponding model framework; the method mainly comprises the following steps:
and step S1, fusing different types of attribute similarity information of the network nodes with structural adjacency information between corresponding network nodes through a first semi-supervised depth model to obtain representations of a plurality of different hidden spaces.
In the step, different types of attribute similarity information of the network nodes are used as input of a first semi-supervised depth model, and then the representation in the hidden space is finely adjusted by utilizing the structural adjacency information of the network through a back propagation algorithm.
The attribute similarity information corresponds to the similarity matrix of the attributes, and the similarity matrix C of the t-th attribute t Ith row and jth column element in (2)
Figure GDA0003740674560000031
Representing network nodes v i Attribute vector of t-th attribute
Figure GDA0003740674560000032
And network node v j Attribute vector of tth attribute
Figure GDA0003740674560000033
The similarity between them.
Structural adjacency information corresponding to adjacencies between pairs of network nodes, specifically each pair of network nodes (v) i ,v j ) The weight of the connection between them represents the adjacency between the network node pairs and also the similarity of the topology of the network node pairs. Structural adjacency information is generally represented by a structural adjacency matrix S, in which the ith row and jth column element S i,j I.e. a network node pair (v) i ,v j ) The adjacency therebetween.
By jointly optimizing the structure and attribute similarity information in the first semi-supervised depth model, not only the highly nonlinear network structure information can be reserved, but also the enhancement effect of the attribute information can be improved.
In the model framework, the step corresponds to a multi-attribute enhanced representation part.
In the embodiment of the invention, the used first semi-supervised depth model, also called single attribute enhanced depth network, is an improvement on a depth autoencoder, and the characterization calculation formula of each layer of hidden space is as follows:
Figure GDA0003740674560000041
wherein the content of the first and second substances,
Figure GDA0003740674560000042
respectively representing network nodes v i Characterization of the t-th attribute information in the hidden space of the 1 st layer and the k-th layer; k denotes the number of network space layers, T denotes the number of attributes,
Figure GDA0003740674560000043
and
Figure GDA0003740674560000044
and
Figure GDA0003740674560000045
and respectively representing the weight and the offset parameter of the network space, wherein the upper mark of the weight and the offset parameter is the serial number of the attribute information, and the lower mark of the weight and the offset parameter is the serial number of the network space layer.
In the embodiment of the invention, the loss function of the first semi-supervised depth model is as follows:
Figure GDA0003740674560000046
wherein the content of the first and second substances,
Figure GDA0003740674560000047
representing a loss function of the unsupervised training process of the first semi-supervised depth model,
Figure GDA0003740674560000048
representing a loss function of the supervised training process of the first semi-supervised depth model,
Figure GDA0003740674560000049
represents L 2 -norm rule item, superscript t being the sequence number of the attribute information; both alpha and beta are hyperparameters, alpha represents
Figure GDA00037406745600000410
Beta represents a weight coefficient of
Figure GDA00037406745600000411
The weight coefficient of (c).
In the first semi-supervised depth model,
Figure GDA00037406745600000412
Figure GDA00037406745600000413
Figure GDA00037406745600000414
where n represents the total number of network nodes, representing a point multiply operation,
Figure GDA00037406745600000415
the decoder stage (decoding stage) representing the model at the first semi-supervised depth corresponds to
Figure GDA00037406745600000416
Is then outputted from the output of (a),
Figure GDA00037406745600000417
is C t Line i of (1), C t A similarity matrix representing the t-th attribute,
Figure GDA00037406745600000418
representing the decoder stage in the first semi-supervised depth model corresponding to C t The matrix of (a) is a matrix of (b),
Figure GDA00037406745600000419
represents H t The number of the ith row of (a),
Figure GDA00037406745600000420
H t represents the adjustment matrix and the adjustment matrix of the device,
Figure GDA00037406745600000421
representation corresponds to matrix C t Row and column i, s i,j The element corresponding to the ith row and jth column of the adjacency matrix S; η and γ are hyper-parameters;
Figure GDA00037406745600000422
and
Figure GDA00037406745600000423
respectively representing network nodes v i And v j The characterization of the tth attribute information in the K layer space of the first semi-supervised depth model;
Figure GDA00037406745600000424
and
Figure GDA00037406745600000425
weight parameters of the k-th layer respectively representing the t-th attribute information and
Figure GDA00037406745600000426
the weight parameter of the corresponding decoder stage.
And step S2, optimizing the structural adjacent information among the network nodes through a second semi-supervised depth model, and obtaining the representation of the network nodes in the hidden space, which is obtained only by using the structural adjacent information.
In the step, the structural adjacency information between the network nodes is used as the input of the second semi-supervised depth model, and then the representation in the hidden space is finely adjusted by using the structural adjacency information between the network nodes through a back propagation algorithm. The step has the effect of enhancing the proportion of the effect of the structural topology information in the final representation, and simultaneously, the step is also used for better processing the condition that the attribute information is too sparse or completely absent.
In the model framework, this step corresponds to the structural characterization part.
In the embodiment of the invention, the second semi-supervised depth model is similar to the first semi-supervised depth model, and the difference is that the attribute similarity matrix C t Replacing with the structural adjacency matrix S and using the same in formula (4)
Figure GDA0003740674560000051
And setting 0.
Network node v obtained in this step for the sake of uniformity of representation i Characterization in hidden space is also denoted
Figure GDA0003740674560000052
Where K is 2]And t is s, s is the network node v i Structural adjacency information with other network nodes. That is to say
Figure GDA0003740674560000053
In (1), T ∈ { [1, T { ]]S, distinguishing according to the value of t
Figure GDA0003740674560000054
Is the result obtained in step S1 or step S2.
Those skilled in the art will appreciate that the first and second terms are used only to distinguish between the two semi-supervised depth models.
And step S3, fusing the representations of the hidden space obtained in the steps S1-S2 to obtain the final network representation.
In the step, the representations of different hidden spaces obtained in the previous two steps are fused, so that a better network representation can be obtained, and the representation is robust to a network with sparsity and inconsistent attributes and has more effective information to a certain extent.
In the model frame, this step corresponds to the bonding layer portion.
In the embodiment of the invention, two ways of fusing the representations of different hidden spaces are splicing and weight summation respectively;
for the splicing of two matrixes, the data of the same row are spliced into one row, so that a new matrix is formed; that is, the tokens of step S1 and step S2 are merged into a final token, denoted as y i
For the weight summation method, the weight values can be obtained by an attention mechanism through training with label training data, and the formula is as follows:
Figure GDA0003740674560000055
wherein, y i As a final network nodev i The characterization of (a) is performed,
Figure GDA0003740674560000056
corresponding to the weight of the hidden space characterization, when T is belonged to [1, T]When t is s, it represents structural adjacency information.
The attention-based weight learning method is defined as follows:
Figure GDA0003740674560000057
wherein, when T is epsilon [1, T]When, G t A weight vector representing the t-th attribute information, wherein G is the value of t ═ s t A weight vector representing the structural adjacency information,
Figure GDA0003740674560000058
is a network node v i The concatenation results of the representations of the different hidden spaces obtained in steps S1 to S2, σ (·) represents a sigmoid function.
The characterization of multiple hidden spaces of each network node obtained by splicing the previous modules, or the characterization y with robustness and discrimination can be obtained by formula (6) i At the same time, such characterization is somewhat more informative.
On the other hand, the embodiment of the invention also provides an optimization algorithm of the model, aiming at realizing the optimization of the first semi-supervised depth model and the second semi-supervised depth model:
1) the first and second semi-supervised depth models are optimized.
The purpose of the optimization is to determine the weight and bias parameters
Figure GDA0003740674560000061
(i.e., parameters)
Figure GDA0003740674560000062
Including first and second semi-supervised depth models
Figure GDA0003740674560000063
And
Figure GDA0003740674560000064
) Minimizing a loss function of the first semi-supervised depth model
Figure GDA0003740674560000065
The optimization process is calculated by the following formula
Figure GDA0003740674560000066
About
Figure GDA0003740674560000067
Partial derivatives of (a):
Figure GDA0003740674560000068
Figure GDA0003740674560000069
Figure GDA00037406745600000610
Figure GDA00037406745600000611
in the above formula, the first and second carbon atoms are,
Figure GDA00037406745600000612
a bias parameter of a k layer representing tth attribute information of a decoding stage of the first semi-supervised depth model;
obtained by the above formula
Figure GDA00037406745600000613
And then adjusting parameters in the first semi-supervised depth model using a back propagation algorithm.
Similarly, the parameters in the second semi-supervised depth model are adjusted by adopting the method.
2) Obtaining the weights in equation (7) using a back propagation algorithm based on the specific task error
Figure GDA00037406745600000614
For classification tasks, the weight vector G for the t-th attribute information is minimized t The objective function of (a) is as follows:
Figure GDA00037406745600000615
wherein the content of the first and second substances,
Figure GDA00037406745600000616
is a collection of network nodes, wherein the total number of network nodes is n, f (-) is a softmax function,
Figure GDA00037406745600000617
is a set of parameter vectors for the classifier,
Figure GDA00037406745600000618
is a network node v i In the tag vector, if v i Belonging to the category p, setting the p-th bit as 1, otherwise, setting the p-th bit as 0;
then, O is calculated class With respect to G t Then using back propagation algorithm to obtain the optimized weight in formula (7)
Figure GDA0003740674560000071
The above method provided by the embodiments of the present invention can be expressed as follows:
inputting:
Figure GDA0003740674560000072
structural adjacency matrix S, hyper-parameters alpha and beta, wherein
Figure GDA0003740674560000073
Represents a collection of network nodes, epsilon represents a collection of edges (edges embody the connection relationships between network nodes),
Figure GDA0003740674560000074
representing a collection of attribute matrices.
Initialization: in a uniform interval [ -1,1 [ ]]Medium random sampling initialization G t And
Figure GDA0003740674560000075
t∈{[1,T],s}。
and (3) outputting: and (5) network characterization.
1: based on
Figure GDA0003740674560000076
Computing attribute similarity matrices
Figure GDA0003740674560000077
2: pre-training single attribute enhanced deep network through deep belief network to obtain initial parameters
Figure GDA0003740674560000078
3: repetition of
4: calculation by the aforementioned step S1 or step S2
Figure GDA0003740674560000079
Figure GDA00037406745600000710
For the characterization of the K-th hidden space,
Figure GDA00037406745600000711
for decoder stage to correspond to
Figure GDA00037406745600000712
An output of (d); t ∈ { [1, T ] here]S } also includes the stepS1 and the result of step S2;
5: updating parameter theta through formulas (2), (8-11) k
6: until convergence
7:
Figure GDA00037406745600000713
8: repeat (R) to
9: updating the parameter G by the equations (6), (7), (12) t And
Figure GDA00037406745600000714
10: until convergence
11: the network characterization is obtained by equation (6).
According to the scheme provided by the embodiment of the invention, when the representation of the network node is learned, the topological structure information and the node attribute information of the network are considered, and meanwhile, the auxiliary function of the attribute information is fully exerted; the scheme is particularly suitable for learning the representation of the network with sparse connection and inconsistent attribute information, and a better network representation learning effect is obtained in the mode; the scheme can provide more accurate support for tasks such as node classification, connection prediction, community discovery and targeted advertising.
Through the description of the above embodiments, it is clear to those skilled in the art that the above embodiments may be implemented by software, or by software plus a necessary general hardware platform. With this understanding, the technical solutions of the embodiments can be embodied in the form of a software product, which can be stored in a non-volatile storage medium (which can be a CD-ROM, a usb disk, a removable hard disk, etc.), and includes several instructions for enabling a computer device (which can be a personal computer, a server, or a network device, etc.) to execute the methods according to the embodiments of the present invention.
The above description is only for the preferred embodiment of the present invention, but the scope of the present invention is not limited thereto, and any changes or substitutions that can be easily conceived by those skilled in the art within the technical scope of the present invention are included in the scope of the present invention. Therefore, the protection scope of the present invention should be subject to the protection scope of the claims.

Claims (6)

1. A method for enhancing network representation learning by using multi-attribute information is applied to a social network and is suitable for node classification, and the method comprises the following steps:
step S1, fusing different types of attribute similarity information of the network nodes with structural adjacency information between corresponding network nodes through a first semi-supervised depth model to obtain a plurality of representations of different hidden spaces, wherein the representation calculation formula of each layer of hidden space is as follows:
Figure FDA0003740674550000011
wherein the content of the first and second substances,
Figure FDA0003740674550000012
respectively representing network nodes v i Characterization of the tth attribute information in the hidden space of the layer 1 and the layer k; k denotes the number of network space layers, T denotes the number of attributes,
Figure FDA0003740674550000013
and
Figure FDA0003740674550000014
Figure FDA0003740674550000015
and
Figure FDA0003740674550000016
respectively representing the weight and the offset parameter of the network space, wherein the upper mark of the weight and the offset parameter is the serial number of the attribute information, and the lower mark of the weight and the offset parameter is the serial number of the network space layer; the attribute information includes: gender, reviews, and user portraits;
s2, optimizing the structural adjacency information among the network nodes through a second semi-supervised depth model to obtain the representation of the network nodes in the hidden space, which is obtained only by using the structural adjacency information;
s3, fusing the representations of the hidden space obtained in the steps S1-S2 to obtain a final network representation;
wherein a loss function of the first semi-supervised depth model is:
Figure FDA0003740674550000017
wherein, the first and the second end of the pipe are connected with each other,
Figure FDA0003740674550000018
a loss function representing an unsupervised training process of the first semi-supervised depth model,
Figure FDA0003740674550000019
representing a loss function of the supervised training process of the first semi-supervised depth model,
Figure FDA00037406745500000110
represents L 2 -norm rule item, superscript t being the sequence number of the attribute information; alpha is a
Figure FDA00037406745500000111
Beta represents
Figure FDA00037406745500000112
The weight coefficient of (c).
2. The method for enhancing network characterization learning by using multi-attribute information as claimed in claim 1, wherein, in the first semi-supervised depth model,
Figure FDA00037406745500000113
Figure FDA00037406745500000114
Figure FDA00037406745500000115
where n represents the total number of network nodes,. represents a point multiply operation; c t A similarity matrix representing the t-th attribute, wherein the i-th row and the j-th column of elements
Figure FDA00037406745500000116
Representing network nodes v i Attribute vector of t-th attribute
Figure FDA00037406745500000117
And network node v j Attribute vector of t-th attribute
Figure FDA0003740674550000021
The similarity between the two or more of the images,
Figure FDA0003740674550000022
is C t The number of the ith row of (a),
Figure FDA0003740674550000023
representing the decoding phase in the first semi-supervised depth model corresponds to
Figure FDA0003740674550000024
Is then outputted from the output of (a),
Figure FDA0003740674550000025
representing the decoding phase in the first semi-supervised depth model corresponding to C t The matrix of (a) is,
Figure FDA0003740674550000026
represents the adjustment matrix H t The number of the ith row of (a),
Figure FDA0003740674550000027
s i,j elements corresponding to the ith row and jth column elements of the adjacency matrix S, η and γ being hyper-parameters;
Figure FDA0003740674550000028
and
Figure FDA0003740674550000029
respectively represent nodes v i And v j The representation of the t-th attribute information in the K-th layer space of the first semi-supervised depth model;
Figure FDA00037406745500000210
respectively representing the weight of the kth layer of the tth attribute information and the weight of the kth layer of the tth attribute information in the decoding stage of the first semi-supervised depth model.
3. The method of claim 2, wherein the second semi-supervised depth model is similar to the first semi-supervised depth model except that the attribute similarity matrix C is used t Is replaced by a structural adjacency matrix S, and
Figure FDA00037406745500000211
in the formula
Figure FDA00037406745500000212
Setting 0; network node v i The characteristics in the hidden space are
Figure FDA00037406745500000213
Where K2, K, t s, s denotes a network node v i Structural adjacency information with other network nodesInformation; that is to say
Figure FDA00037406745500000214
In (1), T ∈ { [1, T { ]]S, distinguishing according to the value of t
Figure FDA00037406745500000215
Is the result obtained in step S1 or step S2.
4. The method for enhancing network characterization learning by using multi-attribute information according to claim 1 or 3, wherein there are two ways of fusing the characterizations of the hidden space in step S3, which are splicing and weight summation respectively;
for the weight summation method, the weight values are obtained by training with label training data through an attention mechanism, and the formula is expressed as follows:
Figure FDA00037406745500000216
wherein, y i For the final network node v i The characterization of (a) is performed,
Figure FDA00037406745500000217
corresponding to the weight of the hidden space characterization, when T is belonged to [1, T]When t is s, it represents structural adjacency information;
the attention-based weight learning method is defined as follows:
Figure FDA00037406745500000218
wherein, when T is ∈ [1, T]When, G t G when t is s, a weight vector representing the t-th attribute information t A weight vector representing the structural adjacency information,
Figure FDA00037406745500000219
is a network node v i The concatenation results of the representations of the different hidden spaces obtained in steps S1 to S2, σ (·) represents a sigmoid function.
5. The method of claim 1, further comprising: optimizing the first and second semi-supervised depth models;
the purpose of the optimization is to determine the weight and bias parameters
Figure FDA00037406745500000220
Minimizing a loss function of the first semi-supervised depth model
Figure FDA00037406745500000221
The optimization process is calculated by the following formula
Figure FDA00037406745500000222
About
Figure FDA00037406745500000223
Partial derivatives of (a):
Figure FDA0003740674550000031
Figure FDA0003740674550000032
Figure FDA0003740674550000033
Figure FDA0003740674550000034
in the above-mentioned formula, the compound has the following structure,
Figure FDA0003740674550000035
a bias parameter of a k layer representing tth attribute information of a decoding stage of the first semi-supervised depth model;
obtained by the above formula
Figure FDA0003740674550000036
Then, adjusting parameters in the first half-supervised depth model by using a back propagation algorithm;
the adjustment of the parameters in the second semi-supervised depth model is performed in a similar manner.
6. The method of claim 4, further comprising: obtaining weights using back propagation algorithm based on specific task errors
Figure FDA0003740674550000037
For classification tasks, the weight vector G for the t-th attribute information is minimized t The objective function of (2) is as follows:
Figure FDA0003740674550000038
wherein the content of the first and second substances,
Figure FDA0003740674550000039
is a collection of network nodes, wherein the total number of network nodes is n, f (-) is a softmax function,
Figure FDA00037406745500000310
is a set of parameter vectors for the classifier,
Figure FDA00037406745500000311
is a network node v i In the tag vector, if v i Belonging to the category p, setting the p-th bit to be 1, otherwise, setting the p-th bit to be 0;
then, O is calculated class With respect to G t Partial derivatives of (a) are weighted by a back propagation algorithm
Figure FDA00037406745500000312
CN201810820414.7A 2018-07-24 2018-07-24 Method for enhancing network representation learning by utilizing multi-attribute information Active CN109117943B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201810820414.7A CN109117943B (en) 2018-07-24 2018-07-24 Method for enhancing network representation learning by utilizing multi-attribute information

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201810820414.7A CN109117943B (en) 2018-07-24 2018-07-24 Method for enhancing network representation learning by utilizing multi-attribute information

Publications (2)

Publication Number Publication Date
CN109117943A CN109117943A (en) 2019-01-01
CN109117943B true CN109117943B (en) 2022-09-30

Family

ID=64863053

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201810820414.7A Active CN109117943B (en) 2018-07-24 2018-07-24 Method for enhancing network representation learning by utilizing multi-attribute information

Country Status (1)

Country Link
CN (1) CN109117943B (en)

Families Citing this family (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110083791B (en) * 2019-05-05 2020-04-24 北京三快在线科技有限公司 Target group detection method and device, computer equipment and storage medium
CN110209825B (en) * 2019-06-17 2021-02-12 大连海事大学 Rapid network characterization learning method based on width learning system
CN110796192B (en) * 2019-10-29 2023-03-28 深圳大学 Image classification method and device based on Internet social contact system
CN111831758B (en) * 2020-08-21 2022-09-16 安徽大学 Node classification method and device based on rapid hierarchical attribute network representation learning
CN112100514B (en) * 2020-08-31 2021-10-26 浙江工业大学 Friend recommendation method based on global attention mechanism representation learning

Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2015180368A1 (en) * 2014-05-27 2015-12-03 江苏大学 Variable factor decomposition method for semi-supervised speech features
EP3273387A1 (en) * 2016-07-19 2018-01-24 Siemens Healthcare GmbH Medical image segmentation with a multi-task neural network system

Family Cites Families (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN103246672B (en) * 2012-02-09 2016-06-08 中国科学技术大学 User is carried out method and the device of personalized recommendation
CN106157043B (en) * 2015-03-24 2021-08-17 联想(北京)有限公司 Processing method of recommended object and electronic equipment
US9813287B2 (en) * 2015-07-30 2017-11-07 Huawei Technologies Co., Ltd. Connectivity-aware virtual network embedding
CN106227991A (en) * 2016-07-14 2016-12-14 西南大学 A kind of corporations method for digging
US11475290B2 (en) * 2016-12-30 2022-10-18 Google Llc Structured machine learning for improved whole-structure relevance of informational displays
CN107622307A (en) * 2017-09-11 2018-01-23 浙江工业大学 A kind of Undirected networks based on deep learning connect side right weight Forecasting Methodology

Patent Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2015180368A1 (en) * 2014-05-27 2015-12-03 江苏大学 Variable factor decomposition method for semi-supervised speech features
EP3273387A1 (en) * 2016-07-19 2018-01-24 Siemens Healthcare GmbH Medical image segmentation with a multi-task neural network system

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
基于链接模型的主动半监督社区发现方法;柴变芳等;《计算机应用》;20171110(第11期);全文 *

Also Published As

Publication number Publication date
CN109117943A (en) 2019-01-01

Similar Documents

Publication Publication Date Title
CN109117943B (en) Method for enhancing network representation learning by utilizing multi-attribute information
Zhang et al. A cross-domain recommender system with kernel-induced knowledge transfer for overlapping entities
Liu et al. Shifu2: A network representation learning based model for advisor-advisee relationship mining
Wang et al. Evolutionary extreme learning machine ensembles with size control
CN111737535B (en) Network characterization learning method based on element structure and graph neural network
CN110880019B (en) Method for adaptively training target domain classification model through unsupervised domain
CN108596774A (en) Socialization information recommendation algorithm based on profound internet startup disk feature and system
Tabouy et al. Variational inference for stochastic block models from sampled data
CN111126488A (en) Image identification method based on double attention
CN111737551A (en) Dark network cable detection method based on special-pattern attention neural network
CN112925977A (en) Recommendation method based on self-supervision graph representation learning
CN114511737B (en) Training method of image recognition domain generalization model
CN112381179A (en) Heterogeneous graph classification method based on double-layer attention mechanism
Li et al. Social recommendation using Euclidean embedding
He et al. ACTL: Adaptive codebook transfer learning for cross-domain recommendation
CN112256971A (en) Sequence recommendation method and computer-readable storage medium
Yu et al. Deep metric learning with dynamic margin hard sampling loss for face verification
Shi et al. EBNAS: Efficient binary network design for image classification via neural architecture search
Xue et al. Self-distribution binary neural networks
CN112598089B (en) Image sample screening method, device, equipment and medium
Wu et al. Fusing hybrid attentive network with self-supervised dual-channel heterogeneous graph for knowledge tracing
Hao et al. Deep graph clustering with enhanced feature representations for community detection
Wang et al. Subspace prototype learning for few-Shot remote sensing scene classification
Liu et al. TCD-CF: Triple cross-domain collaborative filtering recommendation
Zhao et al. Graph pooling via Dual-view Multi-level Infomax

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant