CN109189936B - Label semantic learning method based on network structure and semantic correlation measurement - Google Patents

Label semantic learning method based on network structure and semantic correlation measurement Download PDF

Info

Publication number
CN109189936B
CN109189936B CN201810914904.3A CN201810914904A CN109189936B CN 109189936 B CN109189936 B CN 109189936B CN 201810914904 A CN201810914904 A CN 201810914904A CN 109189936 B CN109189936 B CN 109189936B
Authority
CN
China
Prior art keywords
network
label
text
semantic
edge
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN201810914904.3A
Other languages
Chinese (zh)
Other versions
CN109189936A (en
Inventor
王嫄
杨巨成
李政
赵婷婷
陈亚瑞
赵青
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Beijing Contention Technology Co ltd
Original Assignee
Tianjin University of Science and Technology
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Tianjin University of Science and Technology filed Critical Tianjin University of Science and Technology
Priority to CN201810914904.3A priority Critical patent/CN109189936B/en
Publication of CN109189936A publication Critical patent/CN109189936A/en
Application granted granted Critical
Publication of CN109189936B publication Critical patent/CN109189936B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F40/00Handling natural language data
    • G06F40/30Semantic analysis
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F2216/00Indexing scheme relating to additional aspects of information retrieval not explicitly covered by G06F16/00 and subgroups
    • G06F2216/03Data mining

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Health & Medical Sciences (AREA)
  • Artificial Intelligence (AREA)
  • Audiology, Speech & Language Pathology (AREA)
  • Computational Linguistics (AREA)
  • General Health & Medical Sciences (AREA)
  • Physics & Mathematics (AREA)
  • General Engineering & Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Machine Translation (AREA)

Abstract

The invention relates to a label semantic learning method based on network structure and semantic correlation measurement, which comprises the steps of initializing a label network based on user behavior facts to obtain a fact label network G; label network G after constructing a protocol according to fact label network GR(ii) a On-tag network GRConstructing label network G based on random walk strategy by applying improved random walk strategyC(ii) a Tag network G constructed based on tag-related text informationT(ii) a To label network GCLabel network GTAnd (4) carrying out normalization processing, and learning the semantic vector representation of the label by a random walk strategy and a word vector learning method. The invention has reasonable design, not only fully utilizes the network topological structure, but also considers the relevant text information contained in the node, can express and learn the label semantic vector which is easy to operate, high in confidence coefficient, sufficient in expression and low in noise from the topological structure and the text in a shorter time, and can be widely applied to label network learning and label semantic learning of a text set containing labels.

Description

Label semantic learning method based on network structure and semantic correlation measurement
Technical Field
The invention belongs to the technical field of network representation learning, and particularly relates to a label semantic learning method based on a network structure and semantic correlation measurement.
Background
In the network representation learning technology, text semantic learning is mainly a feature representation of a text, that is, a target text object (word, sentence, segment, and piece) is represented in a numerical value (single value, vector, or matrix). In view of the requirements of computing and Semantic modeling applications, models commonly used at present are latent Semantic analysis lsa (latent Semantic analysis) based on matrix singular value decomposition, latent Dirichlet distribution lda (latent Dirichlet allocation) based on probability model, word vector representation model nnlm (neural Network Language model) based on neural Network solution, word2vec, and the like, and these models are mainly used for long texts. In the face of short texts with high sparseness and high noise, researchers propose a method for expanding original short text corpora by using external corpora, such as WordNet, Mesh, Wikipedia and the like, or aggregating short texts according to a certain rule to form 'pseudo documents', however, the two methods have obvious disadvantages, the former is limited by an external corpus, and the latter is limited by an artificial rule. Through the user behavior analysis of the short text, the fact that a user expands the information extension of the short text by using a large number of labels is found, and researchers use the labels to provide methods such as Tag-LDA, TWTM (Tag-Weighted Topic Model), TWDA (Tag-Weighted dictionary Allocation), Labeled LDA, MB-LDA and HGTM (Hashtag-Graph Based Topic Model) for theme semantic modeling of the short text containing the labels, so that a good effect is obtained. At the same time, the model also outputs a vector representation of the label as a byproduct.
In the network research, classification, clustering, abnormal point detection and link prediction need to face the sparsity of links in the network, and network nodes are expressed as vectors, so that the problem can be effectively avoided. The important branch in the network representation learning is the vector representation of the learning network node. The conventional methods mainly include Graph decomposition (Graph Factorization) and Laplacian eigenvector-based methods (Laplacian Eigenmaps). Perozzi et al learns a linear sequence in a network topology based on random walk learning, and uses Mikolov Skip-Gram word vector learning method and learning network node representation for reference, however, the method of Deepwalk only learns based on connection between nodes, and does not consider content information possibly contained by the nodes. Yang et al have demonstrated that Deepwalk is substantially equivalent to the transition probability matrix decomposition of an overlay network, and on the basis, a depth Associated Deep Walk model (Text Associated Deep Walk) is proposed by introducing node features to carry out co-decomposition. Tang et al propose a model LINE (Large-scale Information Network Embedding) for learning and modeling a Large Network, but the model only considers nodes of one degree and two degrees, and cuts off a Large amount of Information transmitted on the Network, and the three-step, four-step and even more distant associations between different nodes are important components of the global structure of the Network, and have important significance. Based on this, Cao et al propose GraRep to fully learn the correlation information between nodes by defining different transition matrices, defining different loss functions for different step numbers.
In conclusion, the existing algorithm is difficult to solve the problems that the original label network comes from the unconstrained collaboration fact of the user, the information is high in noise and variable, the theme boundary is not clear or the theme drifts, and the semantic model fails.
Disclosure of Invention
The invention aims to overcome the defects of the prior art and provides a label semantic learning method based on a network structure and semantic correlation measurement.
The technical problem to be solved by the invention is realized by adopting the following technical scheme:
a label semantic learning method based on network structure and semantic relevance measurement comprises the following steps:
step 1, initializing a tag network based on user behavior facts to obtain a fact tag network G;
step 2, constructing a label network G after a protocol is constructed according to the fact label network GR
Step 3, in the label network GRConstructing label network G based on random walk strategy by applying improved random walk strategyC
Step 4, constructing a label network G based on the text information related to the labelT
Step (ii) of5. To label network GCLabel network GTAnd (4) carrying out normalization processing, and learning the semantic vector representation of the label by a random walk strategy and a word vector learning method.
Further, the fact label network G is:
defining a label network G ═ { V, E } based on text internal co-occurrence, wherein V is all labels in the whole text set D; if any label i and label j appear in a text d at the same time, there is an edge between them, which is marked as eij(ii) a Defining a weight g of an edge in the networkijComprises the following steps:
Figure BDA0001762727120000021
wherein D isi,jFor a text collection containing both labels i and j, hdIs the set of tags for document d.
Further, the processing method in step 2 is as follows:
first, consider the random noise associated with semantic nodes: branch subtraction is carried out to remove weak correlation with strong randomness, and effective edges of the network are restrained to reduce noise; let the incidence matrix after pruning be represented as T, then the element T thereinijNamely, the correlation value after pruning the label i and the label j is represented as follows:
Figure BDA0001762727120000022
where δ is the phase threshold at which 20% of the low frequency edge can be truncated, gijIs the weight of the edge in the network;
then, considering that the divergence degrees of the label nodes are different, for an edge in any network, the weight of the network label incidence relation is adjusted according to the difference of the number of the endpoints associated with the two endpoints of the edge, and the label theme association is enhanced:
t′ij=tij*log(N/N(i))*log(N/N(j))。
wherein, t'ijIs GRThe weight of the edge in the network, N is the node contained in the networkThe number of points, N (i), N (j), is the degree of the label i, j on the network graph, and logN/N (i) is a logarithmic expression of the probability p (i) that the current label is associated with the node in the network, (i) N (i)/N reciprocal.
Further, the random walk strategy improved in step 3 is to sample a high-noise complex network structure into a plurality of linear sequences, obtain local microscopic description information of the network by a breadth-first search method, obtain global macroscopic information of the network by a depth-first search method, and obtain a new label association relationship by sliding a window on the linear sequences according to the sampled linear sequences, thereby obtaining a new edge weight.
Further, the processing method in the step 4 comprises:
first, a vocabulary set W having a co-occurrence relationship with a tag i is definediBy wijRepresenting the number of times the label i and the vocabulary j co-occur;
then, w is weighted by the inverse text frequency index IDF, the word tiThe idf value of (a) is calculated as follows:
Figure BDA0001762727120000031
where | D | is the number of documents in the set, | { j: t |, isi∈djIs an inclusive word tiThe number of documents;
secondly, calculate w againijAnd idfjObtaining a text representation vector of the label;
finally, calculating cosine similarity of every two label text expression vectors, and defining
Figure BDA0001762727120000032
A truncation threshold for 80% of the low-frequency edge, and a cosine similarity smaller than
Figure BDA0001762727120000033
The weight of the reserved edge is marked as the value of cosine similarity.
Further, step 5 introduces graph mining in the network random walk processSample preference parameters in tag network GTAnd a label network GCSwitch between the two to benefit from the association of the network structure and the text information; in the semantic vector updating process, linear sequences obtained by random walk of a network are used as sentences, and label semantics are learned based on left and right contexts.
The invention has the advantages and positive effects that:
1. the invention combines text semantic learning and network representation learning technologies, reduces network noise through multi-stage noise reduction, network protocol, random walk and text similarity calculation strategies, obtains a core network, learns vector representation of labels on the basis, not only fully utilizes a network topological structure, but also considers relevant text information contained in nodes, can learn label semantic vectors with easy operation, high confidence, sufficient expression and low noise from the topological structure and the text expression in a short time, and can be widely applied to label network learning and label semantic learning of a text set containing labels and various applications of text mining.
2. The method is reasonable in design, and based on the fact information, the core abstraction of the tag association network is performed, the edges on the graph are sampled and the weight is updated, so that the stable tag network structure is obtained, the model is insensitive to noise interference, and the modeling generalization is improved.
Drawings
FIG. 1 is a schematic diagram of the overall architecture of the present invention.
Detailed Description
The embodiments of the present invention will be described in detail with reference to the accompanying drawings.
The design idea of the invention is as follows: the method is mainly based on a statistical machine learning theory and a text mining technology, and a label network with high credibility is constructed by using behavior data of a user to learn pure semantic representation of labels. Firstly, initializing a tag network by utilizing user behavior facts, secondly reconstructing the tag network based on a reduction technology and an improved random walk technology, thirdly, constructing the tag network by utilizing text similarity related to tags, and finally, learning tag semantic vector representation based on the reconstructed tag network and the tag network based on the text similarity. The network reconstruction can be regarded as an automatic multi-path label strong subject correlation discovery filter, the weak correlation part is filtered, and the influence of noise correlation irrelevant to the subject is reduced, so that the network reconstruction helps to discover and enhance label network representation with more consistent and compact subject semantics to obtain a subject enhanced label semantic vector.
Based on the above design concept, the tag semantic learning method based on the network structure and the semantic relevance metric of the invention, as shown in fig. 1, includes the following steps:
step 1: and initializing the tag network based on the user behavior fact to obtain a fact tag network G based on the user behavior.
In the field of internet text applications, essentially "tags" are a theme. The subject matter herein may be fine grained, such as specific events, times, places, or coarse grained, such as aggregated news, abstract concepts, etc. The appearance of the tags is generated by the text composition of the user, and the relationship between the tags is based on the behavioral facts of the user. Thus, naturally, tags co-occur with text.
The most direct relationship between tags is a co-occurrence relationship in a single text, and a tag network G ═ { V, E } based on text internal co-occurrence is defined, and V is all tags in the whole text set D. If any label i and label j appear in a text d at the same time, there is an edge between them, which is marked as eij. Defining a weight g of an edge in the networkijIs the number of texts containing both tags. Namely, it is
Figure BDA0001762727120000041
Wherein D isi,jFor a text collection containing both labels i and j, hdIs the set of tags for document d. Intuitively, two labels appear in a text with a larger quantity, and the weight of the connecting line of the two labels is larger, so that the semantic association of the two labels is tighter.
By this, a tag network based on user behavior facts is obtained.
In addition to the co-occurrence relationships within the text described above, this step also includes a tag fact network that is constructed indirectly using the user and tag usage relationships or the co-occurrence relationships of external links and tags.
Step 2: label network G after construction of protocol based on fact label network GR
Tag networks constructed directly on the basis of facts are relatively noisy, mainly due to the randomness and variability of the user's usage. Liberty means that users are free to tag text, and there are likely to be situations where multiple words are synonymous, such as "Qinghua university" and "Qinghua"; the difference refers to the understanding of the difference of connotation expressed by the same text, so as to add different labels, for example, different users are "how many people can wear" add labels "padded jacket" or "bikini".
To condense semantics, this step is based on four naive assumptions. (1) The more times two tags are associated, the more closely the subject association between the tags. (2) The more identical neighbors of two tags, the more closely the subject association between the tags. (3) The label set with the same theme has the aggregation effect similar to blocks and communities. (4) Tags that are related to a large number of tags are not closely related to the subject matter of other tags.
Random noise associated with semantic nodes is first considered. And (4) carrying out branch subtraction to remove weak correlation with strong randomness, and constraining the effective edge of the network to reduce noise. Let the incidence matrix after pruning be represented as T, then the element T thereinijNamely the correlation value after the pruning of the label i and the label j.
Figure BDA0001762727120000051
Where δ is the phase threshold that can truncate 20% of the low frequency edges.
Secondly, considering that the divergence degrees of the label nodes are different, for an edge in any network, the weight of the network label incidence relation is adjusted according to the difference of the number of the endpoints associated with the two endpoints of the edge, and the label theme association is enhanced:
t′ij=tij*log(N/N(i))*log(N/N(j))。
here, t'ijIs GRThe weight of the edge in the network, N is the number of nodes included in the network, and N (i) and N (j) are the degrees of labels i and j on the network graph, i.e. the number of edges associated with the vertex. logN/N (i) is a logarithmic expression of the probability p (i) ═ N (i)/N reciprocal that the current tag is associated with the node in the network, which reflects the information amount of the tag i in the network, and may be called Inverse correlation Frequency (IRF), and the larger the value, the better the topic semantic distinction degree is, and the more effective information in the topic semantic association is.
And step 3: tag network G after obtaining the specification in step 2RConstructing label network G based on random walk strategy by applying improved random walk strategyC
This step samples the nodes associated with each node based on an improved random walk strategy. In step 2, we get the weights t 'on the network'ijTwo degree transition weights are defined herein to guide the transition. After the label t walks to the label i, the next unnormalized transition probability of the label j is defined as:
Figure BDA0001762727120000052
wherein
Figure BDA0001762727120000053
For the adjustment factor, the following is defined:
Figure BDA0001762727120000054
dtjthe shortest distance on the graph from label t to label j is defined, and only nodes with the distance of 2 at most are defined here. And the label t is a front node when the user randomly walks from i. When d istjWhen 0, it means t, i.e., j, pii t jRefers to the situation that the label j goes to the label iProbability of going back from tag i to tag j.
According to definition
Figure BDA0001762727120000061
After normalization, n-step random walks are performed m times to obtain m paths with the step length of n, sliding with the window size of s is performed in the paths, an edge is added between the labels co-existing in the window, and the weight is added with 1. After all sliding is finished, the weight is the total number of the label co-occurrence windows, and a new label network G is obtained according to the total numberC
And 4, step 4: tag network G constructed based on tag-related text informationT
This step builds a label network using the text associated with the label, i.e., using the content characteristics of the network nodes themselves. The tags and the text vocabularies of the invention have a co-occurrence relationship, and the more the co-occurrence times, the more the vocabularies can embody the meanings indicated by the tags. Defining vocabulary set W with co-occurrence relation with tag iiBy wijIndicating the number of times the label i co-occurs with the vocabulary j.
For example, a text collection contains three pieces of text, respectively:
"# south African golden brick meeting # [ Times News eye I golden brick meeting: gold city 'gold brick'
'south Africa golden brick meeting # economic community, technology is not separated'
"south Africa gold brick meeting # gold brick country, friendship is often in; the common win is shared by resisting wind and rain. Wish our motherland to grow more strongly, the country is Thai and Min' an. "
Wherein, the word "golden brick" labeled "# south Africa golden brick meeting #" has a weight w of 3.
After text representation, w is weighted by the inverse text frequency index idf (inverse Document frequency). Word tiThe idf value of (a) is calculated as follows:
Figure BDA0001762727120000062
wherein | D | is in the setThe number of documents, | { j: ti∈djIs an inclusive word tiThen calculates wijAnd idfjThe product of (a) and (b) to obtain a text representation vector of the tag.
Calculating the cosine similarity of the text representation vectors of every two labels, and defining similar to the step 2
Figure BDA0001762727120000063
Is a truncation threshold that truncates 80% of the low frequency edges. Removing cosine similarity smaller than
Figure BDA0001762727120000064
The edge of (2). The weight of the reserved edge is marked as the value of the cosine similarity.
And 5: for the label network G obtained in the step 3 and the step 4CLabel network GTAnd (4) carrying out normalization, and learning the semantic vector representation of the label by a random walk strategy and a word vector learning method.
G obtained in step 3CEmbodying label association based on network structure, G obtained in step 4TTag association based on semantic relevance is embodied. This step will fuse GCAnd GTAnd two networks, wherein the semantic representation of the label is learned on the converged network.
Since the weight ranges of the two networks are different, the weights of the two networks are unified to be between 0 and 1. Normalization was performed using a linear function, which is as follows:
Figure BDA0001762727120000071
i.e. the original data is scaled equally, xnormIs a normalized value of x, xminAnd xmaxRespectively the minimum and maximum of the edge weights in the network. The normalized networks are respectively GC-normAnd GT-norm. Then random walk is carried out on two networks to obtain a neighbor sequence of each node, word2vec learning is carried out on the sequence, and a SkipGram method is used for updating the vector of the label nodeAnd (4) showing. The calculation flow of the tag semantic vector Φ is as follows:
Figure BDA0001762727120000072
wherein, Shuffle (-) is used for reordering the label nodes to avoid preference caused by operation sequence. RandomWalk (G)C-normJ, t) is completed in graph GC-normStarting from j, randomly walking t steps to obtain a label node sequence W with the length of tj,RandomWalk(GT-normJ, t) are similar. SkipGram (phi, W)jAnd d) a word vector learning method based on the node sequence WjAnd updating the label semantic vector phi with the dimension d.
It should be emphasized that the embodiments described herein are illustrative rather than restrictive, and thus the present invention is not limited to the embodiments described in the detailed description, but also includes other embodiments that can be derived from the technical solutions of the present invention by those skilled in the art.

Claims (1)

1. A label semantic learning method based on network structure and semantic correlation measurement is characterized in that: the method comprises the following steps:
step 1, initializing a tag network based on user behavior facts to obtain a fact tag network G;
step 2, constructing a label network G after a protocol is constructed according to the fact label network GR
Step 3, in the label network GRConstructing label network G based on random walk strategy by applying improved random walk strategyC
Step 4, constructing a label network G based on the text information related to the labelT
Step 5, label network GCLabel network GTCarrying out normalization processing, and learning the semantic vector representation of the label by a random walk strategy and a word vector learning method;
the appearance of the labels is generated by the text composition of the user, the relationship between the labels is based on the behavior fact of the user, and the labels and the text have a co-occurrence relationship;
the fact label network G is:
defining a label network G ═ { V, E } based on text internal co-occurrence, wherein V is all labels in the whole text set D; if any label i and label j appear in a text d at the same time, there is an edge between them, which is marked as eij(ii) a Defining a weight g of an edge in the networkijComprises the following steps:
Figure FDA0003117152020000011
wherein D isi,jFor a text collection containing both labels i and j, hdA set of tags for document d;
the processing method of the step 2 comprises the following steps:
first, consider the random noise associated with semantic nodes: pruning to remove weak correlation with strong randomness, and constraining the effective edge of the network to reduce noise; let the incidence matrix after pruning be represented as T, then the element T thereinijNamely, the correlation value after pruning the label i and the label j is represented as follows:
Figure FDA0003117152020000012
where δ is the phase threshold at which 20% of the low frequency edge can be truncated, gijIs the weight of the edge in the network;
then, considering that the divergence degrees of the label nodes are different, for an edge in any network, the weight of the network label incidence relation is adjusted according to the difference of the number of the endpoints associated with the two endpoints of the edge, and the label theme association is enhanced:
t′ij=tij*log(N/N(i))*log(N/N(j));
wherein, t'ijIs GRThe weight of the edge in the network, N is the number of nodes contained in the network, N (i), N (j) are the degree of the labels i, j on the network graphLog (N/N (i)) is a logarithmic expression of the probability p (i) ═ N (i)/N reciprocal of the label i associated with the nodes in the network;
the random walk strategy improved in the step 3 is to sample a high-noise complex network structure into a plurality of linear sequences, obtain local microscopic description information of the network by using a width-first search method, obtain global macroscopic information of the network by using a depth-first search method, and obtain a new label incidence relation by sliding a window on the linear sequences according to the sampled linear sequences, thereby obtaining a new edge weight;
the processing method of the step 4 comprises the following steps:
first, a vocabulary set W having a co-occurrence relationship with a tag i is definediBy wijRepresenting the number of times the label i and the vocabulary j co-occur;
then, w is paired with the inverse text frequency index idfijWeighted, word tiThe idf value of (a) is calculated as follows:
Figure FDA0003117152020000021
where | D | is the number of documents in the set, | { j: t |, isi∈djIs an inclusive word tiThe number of documents;
secondly, calculate w againijAnd idfjObtaining a text representation vector of the label;
finally, calculating cosine similarity of every two label text expression vectors, and defining
Figure FDA0003117152020000022
A truncation threshold for 80% of the low-frequency edge, and a cosine similarity smaller than
Figure FDA0003117152020000023
The weight of the reserved edge is marked as a cosine similarity value;
in the step 5, in the random walk process of the network, a graph sampling preference parameter is introduced, and a label network G is providedTAnd label networkGCSwitch between the two to benefit from the association of the network structure and the text information; in the semantic vector updating process, linear sequences obtained by random walk of a network are used as sentences, and label semantics are learned based on left and right contexts.
CN201810914904.3A 2018-08-13 2018-08-13 Label semantic learning method based on network structure and semantic correlation measurement Active CN109189936B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201810914904.3A CN109189936B (en) 2018-08-13 2018-08-13 Label semantic learning method based on network structure and semantic correlation measurement

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201810914904.3A CN109189936B (en) 2018-08-13 2018-08-13 Label semantic learning method based on network structure and semantic correlation measurement

Publications (2)

Publication Number Publication Date
CN109189936A CN109189936A (en) 2019-01-11
CN109189936B true CN109189936B (en) 2021-07-27

Family

ID=64921583

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201810914904.3A Active CN109189936B (en) 2018-08-13 2018-08-13 Label semantic learning method based on network structure and semantic correlation measurement

Country Status (1)

Country Link
CN (1) CN109189936B (en)

Families Citing this family (14)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109948066B (en) * 2019-04-16 2020-12-11 杭州电子科技大学 Interest point recommendation method based on heterogeneous information network
CN110413865A (en) * 2019-08-02 2019-11-05 知者信息技术服务成都有限公司 Semantic expressiveness model and its method based on alternating binary coding device characterization model
CN112580916B (en) * 2019-09-30 2024-05-28 深圳无域科技技术有限公司 Data evaluation method, device, computer equipment and storage medium
CN110851491B (en) * 2019-10-17 2023-06-30 天津大学 Network link prediction method based on multiple semantic influence of multiple neighbor nodes
CN110889001B (en) * 2019-11-25 2021-11-05 浙江财经大学 Big image sampling visualization method based on image representation learning
CN111460118B (en) * 2020-03-26 2023-10-20 聚好看科技股份有限公司 Artificial intelligence conflict semantic recognition method and device
CN111444352A (en) * 2020-03-26 2020-07-24 深圳壹账通智能科技有限公司 Knowledge graph construction method and device based on knowledge node membership
CN111611498B (en) * 2020-04-26 2024-01-02 北京科技大学 Network representation learning method and system based on field internal semantics
CN111723301B (en) * 2020-06-01 2022-05-27 山西大学 Attention relation identification and labeling method based on hierarchical theme preference semantic matrix
CN112131569B (en) * 2020-09-15 2024-01-05 上海交通大学 Risk user prediction method based on graph network random walk
CN112182511B (en) * 2020-11-27 2021-02-19 中国人民解放军国防科技大学 Complex semantic enhanced heterogeneous information network representation learning method and device
CN112529621A (en) * 2020-12-10 2021-03-19 中山大学 Advertisement audience basic attribute estimation method based on heterogeneous graph embedding technology
CN112688813B (en) * 2020-12-24 2022-07-15 中国人民解放军战略支援部队信息工程大学 Routing node importance ordering method and system based on routing characteristics
CN112989209B (en) * 2021-05-10 2021-09-17 腾讯科技(深圳)有限公司 Content recommendation method, device and storage medium

Citations (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN105956197A (en) * 2016-06-15 2016-09-21 杭州量知数据科技有限公司 Social media graph representation model-based social risk event extraction method
CN106055604A (en) * 2016-05-25 2016-10-26 南京大学 Short text topic model mining method based on word network to extend characteristics
CN106874931A (en) * 2016-12-30 2017-06-20 东软集团股份有限公司 User portrait grouping method and device
CN107122455A (en) * 2017-04-26 2017-09-01 中国人民解放军国防科学技术大学 A kind of network user's enhancing method for expressing based on microblogging
CN107291803A (en) * 2017-05-15 2017-10-24 广东工业大学 A kind of network representation method for merging polymorphic type information
WO2018118546A1 (en) * 2016-12-21 2018-06-28 Microsoft Technology Licensing, Llc Systems and methods for an emotionally intelligent chat bot
CN108255809A (en) * 2018-01-10 2018-07-06 北京海存志合科技股份有限公司 Consider the method for calculating the theme corresponding to document of Words similarity

Patent Citations (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN106055604A (en) * 2016-05-25 2016-10-26 南京大学 Short text topic model mining method based on word network to extend characteristics
CN105956197A (en) * 2016-06-15 2016-09-21 杭州量知数据科技有限公司 Social media graph representation model-based social risk event extraction method
WO2018118546A1 (en) * 2016-12-21 2018-06-28 Microsoft Technology Licensing, Llc Systems and methods for an emotionally intelligent chat bot
CN106874931A (en) * 2016-12-30 2017-06-20 东软集团股份有限公司 User portrait grouping method and device
CN107122455A (en) * 2017-04-26 2017-09-01 中国人民解放军国防科学技术大学 A kind of network user's enhancing method for expressing based on microblogging
CN107291803A (en) * 2017-05-15 2017-10-24 广东工业大学 A kind of network representation method for merging polymorphic type information
CN108255809A (en) * 2018-01-10 2018-07-06 北京海存志合科技股份有限公司 Consider the method for calculating the theme corresponding to document of Words similarity

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
一种大规模网络中基于节点结构特征映射的链接预测方法;李志宇,梁循,周小平,张海燕,马跃峰;《计算机学报》;20161031;第39卷(第10期);1947-1964 *

Also Published As

Publication number Publication date
CN109189936A (en) 2019-01-11

Similar Documents

Publication Publication Date Title
CN109189936B (en) Label semantic learning method based on network structure and semantic correlation measurement
CN108052593B (en) Topic keyword extraction method based on topic word vector and network structure
CN110765775A (en) Self-adaptive method for named entity recognition field fusing semantics and label differences
CN105528437B (en) A kind of question answering system construction method extracted based on structured text knowledge
CN109753589A (en) A kind of figure method for visualizing based on figure convolutional network
CN111078833B (en) Text classification method based on neural network
JP2021508866A (en) Promote area- and client-specific application program interface recommendations
CN107688576B (en) Construction and tendency classification method of CNN-SVM model
CN107357899B (en) Short text sentiment analysis method based on sum-product network depth automatic encoder
CN112559684A (en) Keyword extraction and information retrieval method
WO2013049529A1 (en) Method and apparatus for unsupervised learning of multi-resolution user profile from text analysis
Fu et al. Long short-term memory network over rhetorical structure theory for sentence-level sentiment analysis
JP2021508391A (en) Promote area- and client-specific application program interface recommendations
CN113051932B (en) Category detection method for network media event of semantic and knowledge expansion theme model
Sadr et al. Unified topic-based semantic models: a study in computing the semantic relatedness of geographic terms
CN114462392A (en) Short text feature expansion method based on topic relevance and keyword association
CN114692605A (en) Keyword generation method and device fusing syntactic structure information
Saha et al. Sentiment Classification in Bengali News Comments using a hybrid approach with Glove
CN115600582B (en) Controllable text generation method based on pre-training language model
CN116992040A (en) Knowledge graph completion method and system based on conceptual diagram
CN114491076B (en) Data enhancement method, device, equipment and medium based on domain knowledge graph
CN110597982A (en) Short text topic clustering algorithm based on word co-occurrence network
CN111782964B (en) Recommendation method of community posts
CN110377845B (en) Collaborative filtering recommendation method based on interval semi-supervised LDA
Kim Research on Text Classification Based on Deep Neural Network

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant
TR01 Transfer of patent right

Effective date of registration: 20240403

Address after: Room 1518B, Unit 2, 12th Floor, Huizhi Building, No. 9 Xueqing Road, Haidian District, Beijing, 100080

Patentee after: Beijing contention Technology Co.,Ltd.

Country or region after: China

Address before: No.9, 13th Street, economic and Technological Development Zone, Binhai New Area, Tianjin

Patentee before: TIANJIN University OF SCIENCE AND TECHNOLOGY

Country or region before: China

TR01 Transfer of patent right