CN111259142A - Specific target emotion classification method based on attention coding and graph convolution network - Google Patents

Specific target emotion classification method based on attention coding and graph convolution network Download PDF

Info

Publication number
CN111259142A
CN111259142A CN202010036470.9A CN202010036470A CN111259142A CN 111259142 A CN111259142 A CN 111259142A CN 202010036470 A CN202010036470 A CN 202010036470A CN 111259142 A CN111259142 A CN 111259142A
Authority
CN
China
Prior art keywords
context
specific target
coding
attention
preset
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN202010036470.9A
Other languages
Chinese (zh)
Other versions
CN111259142B (en
Inventor
肖路巍
刘一凡
胡晓辉
薛云
古东宏
陈秉良
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Shanghai Airlines Intellectual Property Services Ltd
Nanjing Silicon Intelligence Technology Co Ltd
Original Assignee
South China Normal University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by South China Normal University filed Critical South China Normal University
Priority to CN202010036470.9A priority Critical patent/CN111259142B/en
Publication of CN111259142A publication Critical patent/CN111259142A/en
Application granted granted Critical
Publication of CN111259142B publication Critical patent/CN111259142B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/30Information retrieval; Database structures therefor; File system structures therefor of unstructured textual data
    • G06F16/35Clustering; Classification
    • G06F16/353Clustering; Classification into predefined classes
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/044Recurrent networks, e.g. Hopfield networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Data Mining & Analysis (AREA)
  • General Physics & Mathematics (AREA)
  • General Engineering & Computer Science (AREA)
  • General Health & Medical Sciences (AREA)
  • Artificial Intelligence (AREA)
  • Biophysics (AREA)
  • Evolutionary Computation (AREA)
  • Biomedical Technology (AREA)
  • Molecular Biology (AREA)
  • Computing Systems (AREA)
  • Computational Linguistics (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Mathematical Physics (AREA)
  • Software Systems (AREA)
  • Health & Medical Sciences (AREA)
  • Databases & Information Systems (AREA)
  • Machine Translation (AREA)

Abstract

The invention provides a specific target emotion classification method based on attention coding and graph convolution network, comprising the following steps: obtaining hidden state vectors corresponding to the context and the specific target through a preset bidirectional cyclic neural network model, and respectively carrying out multi-head self-attention coding on the hidden state vectors; extracting a syntactic vector in a syntactic dependency tree corresponding to the context by combining a graph convolution neural network with point-to-point convolution, and carrying out multi-head self-attention coding on the syntactic vector; and then, interactively fusing the syntax information code and the context semantic information code, the syntax information code and the specific target semantic information code by using multi-head interactive attention, splicing the fused result and the context semantic information code to obtain final feature representation, and obtaining the emotion classification result of the specific target based on the feature representation. Compared with the prior art, the method and the device fully consider the relation between the context and the specific target and the syntactic information, and improve the accuracy of emotion classification.

Description

Specific target emotion classification method based on attention coding and graph convolution network
Technical Field
The invention relates to the technical field of natural language processing, in particular to a specific target emotion classification method based on attention coding and graph convolution network.
Background
Emotion analysis is an important task in Natural Language Processing (NLP), and aims to analyze subjective text with emotional colors. The emotion analysis of the specific target belongs to fine-grained emotion analysis, and is different from the traditional emotion analysis, and the purpose of the emotion analysis is mainly to identify the emotion polarity of the specific target in a sentence.
At present, a plurality of methods combining a neural network and an attention mechanism are used for solving the emotion analysis problem of a specific target, and although the methods can overcome the defects of a shallow learning model and distinguish the importance of different words to the emotion analysis task of the specific target, the following problems still exist: on one hand, the prior method can not fully capture the semantic information of the context and does not pay attention to the problems of long-distance information dependence and information parallel computation; on the other hand, the existing methods do not consider the relation between the context and the specific target and the syntactic information, and the syntactical constraint technology is lacked, so that syntactical irrelevant context words can be identified as clues for judging the target emotion classification, and the classification accuracy is reduced.
Disclosure of Invention
In order to overcome the problems in the related art, the embodiment of the invention provides a specific target emotion classification method and device based on attention coding and graph convolution network.
According to a first aspect of the embodiments of the present invention, there is provided a method for classifying a specific target emotion based on attention coding and graph convolution network, including the following steps:
obtaining a word vector corresponding to the context and a word vector corresponding to the specific target;
inputting the word vector corresponding to the context and the word vector corresponding to the specific target into a preset bidirectional recurrent neural network model to obtain a hidden state vector corresponding to the context and a hidden state vector corresponding to the specific target;
extracting a syntactic vector in a syntactic dependency tree corresponding to the context based on a preset graph convolutional neural network combined with point-to-point convolution;
performing multi-head self-attention coding on the hidden state vector corresponding to the context, the hidden state vector corresponding to the specific target and the syntax vector to respectively obtain context semantic information coding, specific target semantic information coding and syntax information coding;
respectively carrying out multi-head interactive attention coding on the context semantic information code and the syntax information code and the specific target semantic information code and the syntax information code to obtain a context-syntax information code and a specific target-syntax information code;
averaging and pooling the context semantic information codes, the context-syntax information codes and the specific target-syntax information codes, and then splicing to obtain a feature representation corresponding to the specific target;
and inputting the feature representation into a preset normalization index function to obtain an emotion classification result of the specific target.
Optionally, the word vector corresponding to the context is used
Figure BDA0002366196980000021
Word vectors corresponding to specific targets
Figure BDA0002366196980000022
Inputting the hidden state vector corresponding to the context and the hidden state vector corresponding to the specific target through a preset bidirectional cyclic neural network model LSTM, wherein the process comprises the following steps:
Figure BDA0002366196980000023
and
Figure BDA0002366196980000024
wherein n represents the dimension of the word vector corresponding to the context, m represents the dimension of the word vector corresponding to the specific target,
Figure BDA0002366196980000025
representing the forward operation process in the preset bidirectional recurrent neural network model LSTM,
Figure BDA0002366196980000026
representing the inverse operation process in a preset bidirectional recurrent neural network model LSTM, HcRepresenting a hidden state vector corresponding to the context, HtRepresenting the hidden state vector corresponding to the context.
Optionally, a function is assigned according to the position and the position weight of the specific target in the context to obtain the position weight corresponding to each word in the context;
obtaining a syntactic dependency tree corresponding to a context;
obtaining an adjacency matrix corresponding to the words in the context according to the syntactic dependency tree, wherein the adjacency matrix reflects the adjacency relation of the words in the context;
inputting the adjacency matrix and the position weight corresponding to each word into a preset graph convolution neural network to obtain an output result of an output layer;
and performing point-by-point convolution on the output result to obtain the syntactic vector.
Optionally, a function is assigned according to the position and the position weight of the specific target in the context to obtain the position weight corresponding to each word in the context; wherein the location weight assignment function F (-) is as follows:
Figure BDA0002366196980000027
τ +1 represents a specific purposeStarting position of target, m represents number of words in specific target, n represents number of words in context, q represents number of words in contextiIndicating the position weight of the ith word in the context.
Optionally, inputting the adjacency matrix, the position weight of each word and the output result of the previous layer into a preset graph convolution operation formula to obtain the output result of the current layer, and repeatedly executing the input operation until the output result of the output layer is obtained; the preset graph convolution operation formula is as follows:
Figure BDA0002366196980000031
Figure BDA0002366196980000032
Figure BDA0002366196980000033
Aijthe value representing the ith row and jth column of the adjacency matrix, A ∈ Rn×nRepresenting the contiguous matrix A as a matrix of n rows and n columns, AijRepresenting the output result of the previous layer of the convolutional neural network of the preset map, qlRepresenting the location weight of the jth word in the context,
Figure BDA0002366196980000034
the l-1 layer output result of the jth word of the convolutional neural network representing the preset graph,
Figure BDA0002366196980000035
l-level output result of j-th word representing preset graph convolutional neural network, diIndicates the depth, W, of the ith word in the syntactic dependency treelRepresenting the weight, bias blRepresenting a bias, RELU () representing an activation function.
Optionally, inputting an output result of an output layer of a preset graph convolution network into a preset point-by-point convolution formula to obtain the syntax vector; wherein the preset point-by-point convolution formula is as follows:
Figure BDA0002366196980000036
PWC(h)=σ(h*Wpwc+bpwc)
hlrepresents the output result of the output layer of the preset graph convolution network,
Figure BDA0002366196980000037
represents a syntactic vector, where σ represents the activation function ReLu, a convolution operation,
Figure BDA0002366196980000038
is the learnable weight of the convolution kernel,
Figure BDA0002366196980000039
is the offset of the convolution kernel.
Optionally, the hidden state vector H corresponding to the context is usedcThe hidden state vector H corresponding to the specific targettAnd the syntactic vector
Figure BDA00023661969800000310
Respectively inputting a preset multi-head attention coding formula to obtain a context semantic information code HcsTarget-specific semantic information encoding HtsAnd syntax information coding Hgs(ii) a Wherein, the preset multi-head attention coding formula is as follows:
Hcs=MHA(Hc,Hc)
Hts=MHA(Ht,Ht)
Figure BDA0002366196980000041
Figure BDA0002366196980000042
oh=Attentionh(k,q)
Attention(k,q)=soft max(fs(k,q))k
Figure BDA0002366196980000043
fs(ki,qj)=tanh([ki;qj]·Watt)
fs(ki,qj) First input vector k ═ k, { representing multi-head attention1,k2,...,knAnd a second input vector q ═ q of multi-head attention1,q2,...,qmSemantic relevance of }, when multi-head self-attention coding is performed, k ≠ q, and when multi-head interactive attention coding is performed, k ≠ q, "; "refers to the concatenation of the vectors,
Figure BDA0002366196980000044
is a weight that can be learned that is,
Figure BDA0002366196980000045
represents 1 line 2dhidMatrix of columns, dhidRepresenting the dimension of the hidden state vector softmax () representing the normalized exponential function, Attentionh(k, q) and ohRepresenting the h-th output result in multi-head attention, h e [1, nhead],
Figure BDA0002366196980000046
Represents a pair ohD undergoing a linear changehidLine dhidA matrix of the columns is formed,
Figure BDA0002366196980000047
Figure BDA0002366196980000048
dhrepresenting the vector dimensions of the multi-headed attention-coding output.
Optionally, the context semantic information is encoded HcsAnd syntax information coding HgsThe specific target semantic information code HtsAnd syntax information coding HgsRespectively inputting a preset multi-head attention coding formula to obtain a context-syntax information code HgtAnd specific target-syntax information encoding Hcg(ii) a Wherein the content of the first and second substances,
Hgt=MHA(Hgs,Hts)
Hcg=MHA(Hgs,Hcs)
Figure BDA0002366196980000049
optionally, encoding H context-syntax informationcgSpecific object-syntax information coding HgtAnd context semantic coding HcsInputting a preset average pooling calculation formula, and splicing output results to obtain a feature expression u corresponding to the specific target; wherein, the preset average pooling calculation formula is as follows:
Figure BDA0002366196980000051
Figure BDA0002366196980000052
Figure BDA0002366196980000053
Figure BDA0002366196980000054
represents an average pooling result of context-syntax information encoding,
Figure BDA0002366196980000055
represents an average pooling result of a specific target-syntax information encoding,
Figure BDA0002366196980000056
representing average pooled results of context semantic coding, feature representation
Figure BDA0002366196980000057
"; "refers to the concatenation of vectors.
Optionally, the feature expression is input into a preset conversion formula, and then a conversion result of the feature expression is input into a preset normalization index function, so as to obtain an emotion classification result of the specific target, where the preset conversion formula and the preset normalization index function are as follows:
Figure BDA0002366196980000058
Figure BDA0002366196980000059
Figure BDA00023661969800000510
in order to be a representation of the feature,
Figure BDA00023661969800000511
and bu∈RcAre learnable weight and bias terms, respectively, y ∈ RcIs the emotion classification result of the specific object, and c represents the classified category.
Compared with the prior art, the embodiment of the invention obtains the hidden state vector corresponding to the context and the hidden state vector corresponding to the specific target by presetting the bidirectional cyclic neural network model, and performs multi-head self-attention coding on the hidden state vector corresponding to the context and the hidden state vector corresponding to the specific target by combining the advantages of multi-head self-attention parallel calculation and long-distance dependence, thereby extracting abundant and sufficient context semantic information and specific target semantic information. And extracting a syntax vector in a syntax dependency tree corresponding to the context through a point-by-point convolution graph convolutional neural network, performing multi-head self-attention coding on the syntax vector to obtain syntax information codes, performing interactive fusion on the syntax information codes and the context semantic information codes, the syntax information codes and the specific target semantic information codes respectively by using multi-head interactive attention, splicing the fused result with the context semantic information codes to obtain final feature representation, fully considering the relation between the context, the specific target and the syntax information by the feature representation, preventing context words irrelevant in the syntax from being identified as clues for judging target emotion classification, and improving the accuracy of emotion classification.
It is to be understood that both the foregoing general description and the following detailed description are exemplary and explanatory only and are not restrictive of the invention, as claimed.
For a better understanding and practice, the invention is described in detail below with reference to the accompanying drawings.
Drawings
FIG. 1 is a flowchart illustrating a method for classifying emotion of a specific target based on attention coding and graph convolution network according to an exemplary embodiment of the present invention;
FIG. 2 is a flowchart illustrating S103 in a method for classifying emotion of a specific target based on attention coding and graph convolution network according to an exemplary embodiment of the present invention;
FIG. 3 is a diagram of a syntactic dependency tree provided in an exemplary embodiment of the present invention;
FIG. 4 is a schematic diagram of a adjacency matrix provided by an exemplary embodiment of the invention;
FIG. 5 is a diagram illustrating an overall structure of a specific target emotion classification model based on attention coding and graph convolution network according to an exemplary embodiment of the present invention;
FIG. 6 is a schematic diagram of a convolutional neural network provided in an exemplary embodiment of the present invention;
FIG. 7 is a schematic structural diagram of a specific target emotion classification apparatus based on attention coding and graph convolution network according to an exemplary embodiment of the present invention;
FIG. 8 is a schematic structural diagram of a specific target emotion classification apparatus based on attention coding and graph convolution network according to an exemplary embodiment of the present invention.
Detailed Description
Reference will now be made in detail to the exemplary embodiments, examples of which are illustrated in the accompanying drawings. When the following description refers to the accompanying drawings, like numbers in different drawings represent the same or similar elements unless otherwise indicated. The embodiments described in the following exemplary embodiments do not represent all embodiments consistent with the present invention. Rather, they are merely examples of apparatus and methods consistent with certain aspects of the invention, as detailed in the appended claims.
The terminology used herein is for the purpose of describing particular embodiments only and is not intended to be limiting of the invention. As used in this specification and the appended claims, the singular forms "a", "an", and "the" are intended to include the plural forms as well, unless the context clearly indicates otherwise. It should also be understood that the term "and/or" as used herein refers to and encompasses any and all possible combinations of one or more of the associated listed items.
It is to be understood that although the terms first, second, third, etc. may be used herein to describe various information, these information should not be limited to these terms. These terms are only used to distinguish one type of information from another. For example, first information may also be referred to as second information, and similarly, second information may also be referred to as first information, without departing from the scope of the present invention. The word "if/if" as used herein may be interpreted as "at … …" or "when … …" or "in response to a determination", depending on the context.
Referring to fig. 1, fig. 1 is a flowchart illustrating a specific target emotion classification method based on attention coding and graph convolution network according to an exemplary embodiment of the present invention, where the method is executed by an emotion classification apparatus and includes the following steps:
s101: and acquiring a word vector corresponding to the context and a word vector corresponding to the specific target.
Word embedding is a digital representation of words by mapping a word into a high-dimensional vector, called a word vector, to achieve a representation of the word.
In the embodiment of the present application, the emotion classification apparatus first determines a context and a specific target within the text, where the context may be a sentence in the text, and the specific target is at least one word in the context, for example: the context is "the price is reusable the service is port", and the specific targets are "price" and "service". And then, the emotion classification equipment converts the context and the specific target into corresponding word vectors through a word embedding tool, if the context comprises n words, the word vectors corresponding to the context are n high-dimensional vectors, and if the specific target comprises m words, the word vectors corresponding to the specific target are m high-dimensional vectors.
The word embedding tool can be GloVe or word2vec and the like, and in the embodiment of the application, word vector conversion is performed on the context and the specific target by adopting the GloVe based on the parallelization processing of the GloVe and the advantage of being beneficial to processing a large data set, so that the word vector corresponding to the context and the word vector corresponding to the specific target are obtained.
S102: and inputting the word vector corresponding to the context and the word vector corresponding to the specific target into a preset bidirectional recurrent neural network model to obtain the hidden state vector corresponding to the context and the hidden state vector corresponding to the specific target.
A Recurrent Neural Network (RNN) is a Recurrent Neural Network in which sequence data is input, recursion is performed in an evolution direction of a sequence, and all nodes (cyclic units) are connected in a chain manner, and common Recurrent Neural networks include a Bidirectional Recurrent Neural Network (Bi-RNN), a Long-Short Term Memory Network (LSTM), a Bidirectional Long-Short Term Memory Network (Bi-LSTM), and the like.
In the embodiment of the application, a Bidirectional recurrent neural network (Bi-RNN) or a Bidirectional Long Short-Term Memory network (Bi-LSTM) may be used as a preset Bidirectional recurrent neural network model to obtain a hidden state vector corresponding to a context and a hidden state vector corresponding to a specific target, and initially obtain semantic information included in the context and the specific target. The hidden layer of the bidirectional recurrent neural network model needs to store two values, one value participates in forward calculation, the other value participates in reverse calculation, namely one recurrent network comprises a forward recurrent neural network and a backward neural network, and the bidirectional recurrent neural network is more suitable for modeling time sequence data and is more beneficial to capturing bidirectional semantic dependence.
In an alternative embodiment, since the front-back order of words in context needs to be considered in the process of classifying specific emotion objects, if Bi-LSTM is adopted, information far away can be transmitted, and the problem of long-term dependence is avoided. For example, the context "i don't feel this painting is well-developed", the forward neural network in the Bi-LSTM model can know that the word "not" is negative to the following "good", so as to find that the emotional polarity of the sentence is depreciation. For another example: "I feel that the hotel room is dirty and not good before", can know that 'not go' is a modification to the dirty degree through the backward neural network in Bi-LSTM. Therefore, by utilizing the Bi-LSTM to preset the bidirectional recurrent neural network model, the obtained hidden state vector corresponding to the context and the hidden state vector corresponding to the specific target can describe the context and the semantic information contained in the specific target more accurately.
Specifically, the word vector corresponding to the context is used
Figure BDA0002366196980000081
Word vectors corresponding to specific targets
Figure BDA0002366196980000082
Inputting the hidden state vector corresponding to the context and the hidden state vector corresponding to the specific target through a preset bidirectional cyclic neural network model LSTM, wherein the process comprises the following steps:
Figure BDA0002366196980000083
and
Figure BDA0002366196980000084
wherein n represents a word corresponding to the contextThe dimension of the vector, m represents the dimension of the word vector corresponding to the particular target,
Figure BDA0002366196980000085
representing the forward operation process in the preset bidirectional recurrent neural network model LSTM,
Figure BDA0002366196980000086
representing the inverse operation process in a preset bidirectional recurrent neural network model LSTM, HcRepresenting a hidden state vector corresponding to the context, HtRepresenting the hidden state vector corresponding to the context.
S103: and extracting the syntactic vectors in the syntactic dependency tree corresponding to the context based on a preset graph convolutional neural network combined with point-by-point convolution.
Semantic Dependency analysis (SDP), also known as Dependency tree, is used to analyze Semantic associations between words in a context and present the Semantic associations in a Dependency structure. The specific process is as follows: (1) the context is segmented, for example: "monkeys like to eat bananas. "the monkey likes to eat banana after word segmentation. "; (2) part-of-speech tagging is performed on each word, for example: monkey/NN liked/VV eaten/VV banana/NN. a/PU; (3) generating a phrase syntax tree by part-of-speech tagging; (4) the phrase syntax tree is converted into a syntactic dependency tree.
Graph convolutional neural networks (GCNs) are used to process data of a graph structure type, i.e. a topological structure, which may also be referred to as a non-euclidean structure, and common graph structures include, for example, social networks, information networks, and the like. The point-by-point convolution is to perform convolution operation again on the output result of the graph convolution neural network so as to better integrate the syntax in the sentence.
In the embodiment of the application, the emotion classification device acquires a syntactic dependency tree corresponding to a context, and acquires an initial syntactic vector in the syntactic dependency tree by using a preset graph convolution neural network, and then performs point-by-point convolution on the initial syntactic vector to obtain a syntactic vector corresponding to the context. The trained graph convolution neural network is used as a preset graph convolution neural network, and the specific training mode is the same as that of the existing neural network.
The syntactic information displayed in the syntactic dependency tree can be extracted by combining the preset graph convolutional neural network of point-by-point convolution to obtain a syntactic vector, so that words irrelevant in syntax are avoided being taken as clues for judging emotion polarity in the follow-up process, and the accuracy of emotion classification of a specific target is improved.
In an optional embodiment, to accurately obtain the syntax vector, step S103 specifically includes steps S1031 to S1035, referring to fig. 2, fig. 2 is a schematic flow diagram of step S103 in the specific target emotion classification method based on attention coding and graph convolution network according to an exemplary embodiment of the present invention, and steps S1031 to S1035 are as follows:
s1031: and obtaining the position weight corresponding to each word in the context according to the position of the specific target in the context and the position weight distribution function.
The importance of each word in the context to the sentiment classification of a particular object varies depending on the location of the particular object in the context. Specifically, the position weight corresponding to each word in the context is obtained according to the position of the specific target in the context and the position weight distribution function. The position weight distribution function can be preset according to different requirements of different specific target emotion classifications. For example: the position weight distribution function can be set as F (a), wherein a is the number of words separated from the nearest specific target between each word in the context, so that different position weights can be obtained according to the number of the separated words.
In an optional embodiment, a position weight word vector corresponding to each word in the context is obtained according to the position of the specific target in the context and a position weight distribution function; wherein the location weight assignment function F (-) is as follows:
Figure BDA0002366196980000091
τ +1 represents the starting position of the specific object, m represents the number of words in the specific object, n represents the number of words in the context,qiindicating the position weight of the ith word in the context.
The importance degree of the words at different positions can be analyzed more accurately through the weight distribution function F (-) and the position weight corresponding to each word in the context can be distributed more reasonably.
S1032: and acquiring a syntactic dependency tree corresponding to the context.
And the emotion classification equipment acquires a syntactic dependency tree corresponding to the context.
In the embodiment of the application, the syntactic dependency tree can be obtained through spaCy. The syntactic dependency tree can vividly embody the dependency relationship of words in the context. Referring to fig. 3, fig. 3 is a diagram illustrating a syntactic dependency tree according to an exemplary embodiment of the present invention. As shown, the context is "I am happy today", and in the constructed syntactic dependency tree, the structural syntactic dependency tree is happy with the root of the syntactic dependency tree, which comprises three branches of "I" (subject), "today" (subject) and "not" (subject), and the branch of "not" is "very" (subject) for further modifying "not".
S1033: and obtaining an adjacency matrix corresponding to the words in the context according to the syntactic dependency tree, wherein the adjacency matrix reflects the adjacency relation of the words in the context.
In the embodiment of the application, the emotion classification device obtains an adjacency matrix corresponding to the words in the context according to the syntactic dependency tree. Where the adjacency matrix reflects the adjacency of words in the context. It should be noted that the word has an adjacency with itself by default.
Referring to fig. 4, fig. 4 is a schematic diagram of an adjacency matrix according to an exemplary embodiment of the present invention, and the adjacency matrix shown in fig. 4 corresponds to the syntactic dependency tree shown in fig. 3. As shown, the values on the diagonal lines are all 1, which indicates that each word has an adjacent relation with itself. The root of the syntactic dependency tree in FIG. 3 is "happy," which includes three branches, "I", "today", and "not", respectively, and thus the value is 1 at the intersection of the row where "happy" is located and the column value where "I", "today", and "not" are located in the corresponding adjacency matrix. The adjacency relation between the words can be accurately and quickly acquired through the adjacency matrix corresponding to the words in the context.
S1034: and inputting the adjacency matrix and the position weight corresponding to each word into a preset graph convolution neural network to obtain an output result of an output layer.
In the embodiment of the application, the emotion classification device inputs the adjacency matrix and the position weight corresponding to each word into a preset graph convolution neural network to obtain an output result of an output layer. The hidden layer of the preset graph convolutional neural network is set to be 1 layer, and the activation function can be set according to actual conditions.
Specifically, the emotion classification device classifies the adjacency matrix, the position weight of each word and the hidden state vector H corresponding to the contextcInputting an input layer of a preset graph convolutional neural network, transmitting an output result of the input layer, an adjacent matrix and the position weight of each word into a hidden layer, transmitting the output result, the adjacent matrix and the position weight of each word to the output layer by the hidden layer, and finally obtaining an output result of the output layer, namely an initial syntax vector in a syntax dependency tree.
In an alternative embodiment, the hidden layer of the pre-defined graph convolution neural network may be a plurality of layers, and the activation function is RELU ().
Specifically, the emotion classification device inputs the adjacency matrix, the position weight of each word and the output result of the previous layer into a preset graph convolution operation formula to obtain the output result of the current layer, and repeatedly executes input operation until the output result of the output layer is obtained; the preset graph convolution operation formula is as follows:
Figure BDA0002366196980000101
Figure BDA0002366196980000111
Figure BDA0002366196980000112
Aijthe value representing the ith row and jth column of the adjacency matrix, A ∈ Rn×nRepresenting the contiguous matrix A as a matrix of n rows and n columns, AijRepresenting the output result of the previous layer of the convolutional neural network of the preset map, qlRepresenting the location weight of the jth word in the context,
Figure BDA0002366196980000113
the l-1 layer output result of the jth word of the convolutional neural network representing the preset graph,
Figure BDA0002366196980000114
l-level output result of j-th word representing preset graph convolutional neural network, diIndicates the depth, W, of the ith word in the syntactic dependency treelRepresenting the weight, bias blRepresenting a bias, RELU () representing an activation function.
S1035: and performing point-by-point convolution on the output result to obtain the syntactic vector.
And the emotion classification equipment performs point-by-point convolution on the output result (namely the initial syntactic vector) to obtain the syntactic vector. Wherein, the point-by-point convolution operation refers to the step-by-step pair of initial syntax vectors
Figure BDA0002366196980000115
Vector of each word in the word
Figure BDA0002366196980000116
A convolution operation is performed to better integrate the syntactic information within each word.
In an optional embodiment, the emotion classification device inputs an output result of an output layer of a preset graph convolution network into a preset point-by-point convolution formula to obtain the syntax vector; wherein the preset point-by-point convolution formula is as follows:
Figure BDA0002366196980000117
PWC(h)=σ(h*Wpwc+bpwc)
hlrepresents the output result of the output layer of the preset graph convolution network,
Figure BDA0002366196980000118
represents a syntactic vector, where σ represents the activation function ReLu, a convolution operation,
Figure BDA0002366196980000119
is the learnable weight of the convolution kernel,
Figure BDA00023661969800001110
is the offset of the convolution kernel.
S104: and carrying out multi-head self-attention coding on the hidden state vector corresponding to the context, the hidden state vector corresponding to the specific target and the syntactic vector to respectively obtain context semantic information coding, specific target semantic information coding and syntactic information coding.
The essence of the attention mechanism comes from the human visual attention mechanism, which is applied to emotion classification in order to enable more attention to be assigned to key words in the classification process. Specifically, a sentence of text can be imagined to be composed of a series of < Key, Value > data pairs, at this time, a certain element Query is given, a weight coefficient of Value corresponding to each Key is obtained by calculating similarity or correlation between the Query and each Key, and after normalization by a softmax function, the weight coefficient and the corresponding Value are subjected to weighted summation to obtain an attention result. In current research, Key and Value are often equal, i.e., Key is Value.
The Multi-head Attention coding (Multi-head Attention) represents that multiple times of Attention coding operation are carried out, each operation represents one head, parameters among the heads are not shared, and finally, the results are spliced and linear transformation is carried out once to obtain a Multi-head coding result.
The multi-head attention coding is further divided into multi-head self-attention coding and multi-head interactive attention coding. The Query and the Key of multi-head self attention are the same, and the Query and the Key of multi-head interactive attention coding are different. For multi-head self-attention coding, it needs to implement the calculation of attention values between each word in a certain sentence text and all words of the sentence text; for multi-headed interactive attention coding, it requires the implementation of the calculation of attention values between each word in a certain sentence of text and all the words of the other text.
In this embodiment of the application, the emotion classification device performs multi-head self-attention coding on the hidden state vector corresponding to the context, the hidden state vector corresponding to the specific target, and the syntax vector to obtain a context semantic information code, a specific target semantic information code, and a syntax information code, respectively.
Specifically, (1) the emotion classification device performs multi-head self-attention coding by taking hidden state vectors corresponding to the context as Query and Key to obtain context semantic information codes; (2) the emotion classification equipment performs multi-head self-attention coding by taking hidden state vectors corresponding to the specific target as Query and Key to obtain semantic information coding of the specific target; (3) and the emotion classification equipment performs multi-head self-attention coding by taking the syntax vector as Query and Key to obtain syntax information coding.
By carrying out multi-head self-attention coding on the hidden state vector corresponding to the context, the hidden state vector corresponding to the specific target and the syntactic vector, richer semantic information and emotional information can be extracted.
In an optional embodiment, the emotion classification device classifies the hidden state vector H corresponding to the contextcThe hidden state vector H corresponding to the specific targettAnd the syntactic vector
Figure BDA0002366196980000121
Respectively inputting a preset multi-head attention coding formula to obtain a context semantic information code HcsTarget-specific semantic information encoding HtsAnd syntax information coding Hgs(ii) a Wherein, the preset multi-head attention coding formula is as follows:
Hcs=MHA(Hc,Hc)
Hts=MHA(Ht,Ht)
Figure BDA0002366196980000122
Figure BDA0002366196980000123
oh=Attentionh(k,q)
Attention(k,q)=soft max(fs(k,q))k
Figure BDA0002366196980000131
fs(ki,qj)=tanh([ki;qj]·Watt)
fs(ki,qj) First input vector k ═ k, { representing multi-head attention1,k2,...,knAnd a second input vector q ═ q of multi-head attention1,q2,...,qmSemantic relevance of }, when multi-head self-attention coding is performed, k ≠ q, and when multi-head interactive attention coding is performed, k ≠ q, "; "refers to the concatenation of the vectors,
Figure BDA0002366196980000132
is a weight that can be learned that is,
Figure BDA0002366196980000133
represents 1 line 2dhidMatrix of columns, dhidRepresenting the dimension of the hidden state vector softmax () representing the normalized exponential function, Attentionh(k, q) and ohRepresenting the h-th output result in multi-head attention, h e [1, nhead],
Figure BDA0002366196980000134
Represents a pair ohD undergoing a linear changehidLine dhidA matrix of the columns is formed,
Figure BDA0002366196980000135
Figure BDA0002366196980000136
dhrepresenting the vector dimensions of the multi-headed attention-coding output.
S105: and respectively carrying out multi-head interactive attention coding on the context semantic information code and the syntax information code and the specific target semantic information code and the syntax information code to obtain a context-syntax information code and a specific target-syntax information code.
And the emotion classification equipment respectively carries out multi-head interactive attention coding on the context semantic information coding and the syntax information coding and the specific target semantic information coding and the syntax information coding to obtain a context-syntax information coding and a specific target-syntax information coding.
Specifically, (1) the emotion classification device performs multi-head interactive attention coding by using syntax information coding as Key and context semantic information coding as Query to obtain context-syntax information coding, so that the syntax information coding and the context semantic coding are interactively fused, and close connection between the syntax information and the context is fully considered. (2) The emotion classification equipment carries out multi-head interactive attention coding by taking the syntax information code as Key and the specific target semantic information code as Query to obtain a specific target-syntax information code, so that the syntax information code and the specific target semantic information code are interactively fused, and the close connection between the syntax information and the specific target is fully considered.
In an alternative embodiment, the emotion classification device encodes the context semantic information HcsAnd syntax information coding HgsThe specific target semantic information code HtsAnd syntax information coding HgsRespectively inputting a preset multi-head attention coding formula to obtain a context-syntax information code HgtAnd specific target-syntax information encoding Hcg(ii) a Wherein the content of the first and second substances,
Hgt=MHA(Hgs,Hts)
Hcg=MHA(Hgs,Hcs)
Figure BDA0002366196980000141
s106: and averaging and pooling the context semantic information codes, the context-syntax information codes and the specific target-syntax information codes, and then splicing to obtain the feature representation corresponding to the specific target.
And the emotion classification equipment averages and pools the context semantic information codes, the context-syntax information codes and the specific target-syntax information codes and then splices the context semantic information codes, the context-syntax information codes and the specific target-syntax information codes to obtain the characteristic representation corresponding to the specific target. Wherein, the average pooling operation is to average the values of the same dimension, and the splicing operation is to splice the vectors end to end, such as vectors [1,1], [2,2], [3,3], to obtain [1,1,2,2,3,3 ].
In an alternative embodiment, the emotion classification device encodes the context-syntax information HcgSpecific object-syntax information coding HgtAnd context semantic coding HcsInputting a preset average pooling calculation formula, and splicing output results to obtain a feature expression u corresponding to the specific target; wherein, the preset average pooling calculation formula is as follows:
Figure BDA0002366196980000142
Figure BDA0002366196980000143
Figure BDA0002366196980000144
Figure BDA0002366196980000145
represents an average pooling result of context-syntax information encoding,
Figure BDA0002366196980000146
represents an average pooling result of a specific target-syntax information encoding,
Figure BDA0002366196980000147
representing average pooled results of context semantic coding, feature representation
Figure BDA0002366196980000148
"; "refers to the concatenation of vectors.
S107: and inputting the feature representation into a preset normalization index function to obtain an emotion classification result of the specific target.
And the emotion classification equipment inputs the feature representation into a preset normalized index function to obtain an emotion classification result of the specific target. The preset normalization index function is a softmax () function, and probability distribution of emotion polarities under different specific targets is obtained through the preset normalization index function, so that emotion classification results of the specific targets are obtained.
In an optional embodiment, the emotion classification device inputs the feature representation into a preset conversion formula, and then inputs a conversion result of the feature representation into a preset normalization index function to obtain an emotion classification result of the specific target, where the preset conversion formula and the preset normalization index function are as follows:
Figure BDA0002366196980000149
Figure BDA0002366196980000151
Figure BDA0002366196980000152
in order to be a representation of the feature,
Figure BDA0002366196980000153
and bu∈RcAre the learnable weight and bias term, respectively, y ∈RcIs the emotion classification result of the specific object, and c represents the classified category.
Referring to fig. 5 and fig. 6, fig. 5 is a schematic diagram illustrating an overall structure of a specific target emotion classification model based on attention coding and a graph convolution network according to an exemplary embodiment of the present invention, and fig. 6 is a schematic diagram illustrating a graph convolution neural network according to an exemplary embodiment of the present invention. The specific target emotion classification model (hereinafter referred to as an AEGCN model) based on attention coding and graph volume network corresponds to the specific target emotion classification method based on attention coding and graph volume network proposed in the embodiment of the present application, for example: steps S101 to S107. Specifically, the model firstly obtains a hidden state vector corresponding to a context and a hidden state vector corresponding to a specific target through a preset bidirectional cyclic neural network model, then performs multi-head self-attention coding on the hidden state vector corresponding to the context and the hidden state vector corresponding to the specific target by combining the advantages of multi-head self-attention parallel computation and long-distance dependence, and extracts abundant and sufficient context semantic information and specific target semantic information. And extracting a syntax vector in a syntax dependency tree corresponding to the context by combining a point-by-point convolution graph convolutional neural network, performing multi-head self-attention coding on the syntax vector to obtain syntax information codes, performing interactive fusion on the syntax information codes, the context semantic information codes, the syntax information codes and the specific target semantic information codes respectively by using multi-head interactive attention, splicing the fused result with the context semantic information codes to obtain final feature representation, fully considering the relation between the context, the specific target and the syntax information by the feature representation, preventing context words irrelevant in syntax from being identified as clues for judging target emotion classification, and improving the accuracy of emotion classification.
The following experimental demonstration is performed on the specific target emotion classification method based on attention coding and graph convolution network, which is provided by the embodiment of the application, and the demonstration process is as follows:
(1) five data sets TWITTER, REST14 and LAP14(SemEval 2014task4), REST15(SemEval 2015 task 12), REST16(SemEval 2016 task 5) were chosen.
Wherein the TWITTER data set was originally created by Tang et al, and contains tweets from social software TWITTER, which collectively include 6940 comments, each with a particular target marked therein and the emotional polarity of the particular target.
The SemEval-2014 Task4 data set is mainly used for fine-grained sentiment analysis and comprises LAP14 and REST14, the data set of each field is divided into training data, verification data (separated from the training data) and test data, and comprises 7794 comments, and each comment has a specific target marked therein and the sentiment polarity of the specific target.
The SemEval-2015 task 12 data set is mainly used for fine-grained emotion analysis and comprises REST15, the data set of each field is divided into training data, verification data (separated from the training data) and test data, and a total of 1746 comments, each comment has a specific target marked therein and the emotional polarity of the specific target.
SemEval-2016 task 5, a data set mainly used for fine-grained sentiment analysis, comprising REST16, wherein the data set of each field is divided into training data, verification data (separated from the training data) and test data, and comprises 2454 comments in total, and each comment has a specific target marked therein and the sentiment polarity of the specific target.
(2) Preprocessing the data in the dataset, specifically, initializing in a GloVe tool, converts each word in the data into a high-dimensional vector with dimension 300. And initializing and setting the weight of the model used in the experiment by using the uniform distribution.
(3) And (4) constructing a graph convolution neural network structure, wherein the dimension of all hidden layers is 300. Adam is an optimizer with a learning rate of 0.001. L is2The weight of the regularization term is set to 0.00001. Meanwhile, to prevent overfitting, the parameter dropout rate is 0.5. The coefficient of the parameter batch size is 32. The number of heads in the multi-head attention is 3, and the number of GCN layers is set to 2.
(4) And comparing the experimental results.
The invention selects Accuracy and Macro-Averaged F1 as evaluation indexes. Wherein, Accuracy is a two-classification evaluation index, the calculation mode is the ratio of the number of correctly classified sentences to the total number of sentences, Macro-Averaged F1 is a multi-classification evaluation index, and the calculation mode is as follows:
Figure BDA0002366196980000161
Figure BDA0002366196980000162
Figure BDA0002366196980000163
Figure BDA0002366196980000164
Figure BDA0002366196980000165
wherein, TPiRefers to the True Positive of the classification i, which refers to the number of sentences predicted classified as i and truly classified as i, FPiFalse Positive for classification i, number of sentences predicted to be classified as i but not really classified as i, TNiTrue Negative referring to classification i, where True Negative refers to the number of sentences predicted to be classified as not i and truly classified as not i, FNiFalse Negative referring to the classification i, False Negative referring to the number of sentences predicted to be classified as not i but actually i, and n is the total number of sentences.
Referring to table 1 below, it can be seen from table 1 that the performance of the specific target emotion classification method based on attention coding and graph convolution network proposed in the present application is superior to that of the conventional machine learning method. The SVM models in the table are classified by using a support vector machine, and rely on a large amount of artificial feature extraction. The specific target emotion classification method based on attention coding and graph convolution network provided by the application has no artificial feature extraction, and the accuracy rates of Twitter, lap14 and restaurant data sets are respectively 9.76%, 5.42% and 0.88% higher than that of SVM. The specific target emotion classification method based on attention coding and graph convolution network is more suitable for research of specific target emotion analysis.
The attention coding and graph convolution network-based specific target emotion classification method uses the bidirectional LSTM and combines the multi-head attention mechanism to encode the semantic, and has better effect compared with a standard multi-attention mechanism method and a method for performing semantic encoding only by using the multi-head attention mechanism. Taking MemNet as an example, the accuracy and F1 values were lower on the five data sets than in the method herein. Taking the AEN as an example, only one index (REST14, F1) was low (0.82%) in the present method among the three data sets (TWITTER, LAP14, and REST 14).
Due to the combination of the syntactic information, compared with a method without considering the syntactic information, the specific target emotion classification method based on attention coding and graph convolution network provided by the application has better effect. Taking AOA as an example, the accuracy of REST16 alone among the accuracy and F1 values on five data sets was slightly higher (0.11% higher) than the method proposed in the present application. In addition, the IAN accuracy rate and the F1 value on five data sets are lower than those of the method, and the method provided by the application is better than TNet-LF in the performance of four data sets.
The GCN combined with point-by-point convolution provided by the invention is superior to the GCN combined with aspect-specific masking layer. The ASGCN model enforces an apect-specific masking layer on top of the GCN to obtain a representation of a particular object that incorporates syntactic information. But this loses the contextual representation with the syntactic information. Of the accuracy and the F1 value of the five data sets, the method only has the three indexes of the accuracy of F1 of REST14, the accuracy of F1 of REST15 and the accuracy of REST16 which are lower than (0.87%, 1.02% and 1.60% of respecitvely) ASGCN, and has the other seven indexes which are better than the ASGCN, so that the effectiveness of the method provided by the application is proved.
Figure 1
TABLE 1
As above, the ablation study was performed for the specific target emotion classification method based on attention coding and graph convolution network proposed in the present application, and the results are shown in table 2.
First, there were three datasets (LAP14, REST14, and REST15) that exhibited a downslide after the GCN was removed, but TWITTER and REST16 performed better. Since the sentence comparison spoken in the TWITTER dataset and the TNet-LF method observed in combination with the above experimental results performed best on both indices of the dataset REST16, it can be concluded that the TWITTER dataset and the REST16 dataset are not very sensitive to syntactic information.
Secondly, without multi-head self-attention, the TWITTER dataset does not perform well, but the REST15 dataset results are rising, which reflects that the TWITTER dataset is very dependent on semantic information and that the application of the invention to multi-head self-attention mechanism can extract rich semantic information well, and it can be inferred that the REST15 dataset is more sensitive to syntactic information than other datasets.
Finally, it can be seen that if multi-head interaction attention is removed, the present invention does not perform well on five data sets, which also indicates that multi-head interaction attention is very important to the present invention, and that interaction of syntactic and semantic information is very important for data sets like LAP14, REST14, and REST 15.
Experimental results of ablation studies show that each step in the specific target emotion classification method based on attention coding and graph convolution network proposed in the embodiments of the present application is indispensable and effective.
Figure BDA0002366196980000181
TABLE 2
Referring to fig. 7, fig. 7 is a schematic structural diagram of a specific target emotion classification apparatus based on attention coding and graph convolution network according to an exemplary embodiment of the present invention. The units are included for executing the steps in the embodiments corresponding to fig. 1 and fig. 2, and refer to the related description in the embodiments corresponding to fig. 1 and fig. 2. For convenience of explanation, only the portions related to the present embodiment are shown. Referring to fig. 7, the specific target emotion classification apparatus 7 based on attention coding and graph convolution network includes:
an obtaining unit 71, configured to obtain a word vector corresponding to a context and a word vector corresponding to a specific target;
the first neural network unit 72 is configured to input the word vector corresponding to the context and the word vector corresponding to the specific target into a preset bidirectional recurrent neural network model, so as to obtain a hidden state vector corresponding to the context and a hidden state vector corresponding to the specific target;
a second neural network unit 73, configured to extract a syntactic vector in the syntactic dependency tree corresponding to the context based on a preset graph convolutional neural network combining point-by-point convolution;
a first encoding unit 74, configured to perform multi-head self-attention encoding on the hidden state vector corresponding to the context, the hidden state vector corresponding to the specific target, and the syntax vector, so as to obtain a context semantic information code, a specific target semantic information code, and a syntax information code, respectively;
a second encoding unit 75, configured to perform multi-head interactive attention encoding on the context semantic information encoding and syntax information encoding, and the specific target semantic information encoding and syntax information encoding, respectively, to obtain a context-syntax information encoding and a specific target-syntax information encoding;
a splicing unit 76, configured to splice the context semantic information codes, the context-syntax information codes, and the specific target-syntax information codes after averaging and pooling, so as to obtain a feature representation corresponding to the specific target;
and a classification unit 77, configured to input the feature representation into a preset normalized exponential function, so as to obtain an emotion classification result of the specific target.
Optionally, the second neural network unit 73 includes:
a position weight assigning unit 731, configured to assign a function according to a position and a position weight of a specific target in a context, and obtain a position weight corresponding to each word in the context;
a first obtaining unit 732, configured to obtain a syntactic dependency tree corresponding to a context;
a second obtaining unit 733, configured to obtain an adjacency matrix corresponding to a word in the context according to the syntactic dependency tree, where the adjacency matrix reflects an adjacency relationship of the word in the context;
a third neural network unit 734, configured to input the adjacency matrix and the position weight corresponding to each word into a preset graph convolution neural network, so as to obtain an output result of an output layer;
and a point-by-point convolution unit 735, configured to perform point-by-point convolution on the output result to obtain the syntax vector.
Referring to fig. 8, fig. 8 is a schematic diagram of a specific target emotion classification apparatus based on attention coding and graph convolution network according to an exemplary embodiment of the present invention. As shown in fig. 8, the specific target emotion classification device 8 based on attention coding and graph convolution network of this embodiment includes: a processor 80, a memory 81, and a computer program 82 stored in said memory 81 and operable on said processor 80, such as a specific target emotion classification program based on attention coding and graph volume networks. The processor 80, when executing the computer program 82, implements the steps in each of the above embodiments of the method for classifying specific target emotion based on attention coding and graph convolution network, such as the steps S101 to S107 shown in fig. 1. Alternatively, the processor 80, when executing the computer program 82, implements the functions of the modules/units in the above-described device embodiments, such as the functions of the units 71 to 77 shown in fig. 7.
Illustratively, the computer program 82 may be partitioned into one or more modules/units that are stored in the memory 81 and executed by the processor 80 to implement the present invention. The one or more modules/units may be a series of computer program instruction segments capable of performing specific functions for describing the execution of the computer program 82 in the specific target emotion classification device 8 based on attention coding and graph convolution network. For example, the computer program 82 may be partitioned into an acquisition unit, a first neural network unit, a second neural network unit, a first coding unit, a second coding unit, a splicing unit, and a classification unit, each unit functioning as follows:
the acquiring unit is used for acquiring word vectors corresponding to the context and word vectors corresponding to the specific target;
the first neural network unit is used for inputting the word vector corresponding to the context and the word vector corresponding to the specific target into a preset bidirectional cyclic neural network model to obtain a hidden state vector corresponding to the context and a hidden state vector corresponding to the specific target;
the second neural network unit is used for extracting a syntactic vector in the syntactic dependency tree corresponding to the context based on a preset graph convolutional neural network combined with point-by-point convolution;
the first coding unit is used for carrying out multi-head self-attention coding on the hidden state vector corresponding to the context, the hidden state vector corresponding to the specific target and the syntax vector to respectively obtain context semantic information coding, specific target semantic information coding and syntax information coding;
the second coding unit is used for respectively carrying out multi-head interactive attention coding on the context semantic information coding and the syntax information coding and the specific target semantic information coding and the syntax information coding to obtain a context-syntax information coding and a specific target-syntax information coding;
the splicing unit is used for splicing the context semantic information codes, the context-syntax information codes and the specific target-syntax information codes after averaging and pooling to obtain the feature representation corresponding to the specific target;
and the classification unit is used for inputting the feature representation into a preset normalized index function to obtain an emotion classification result of the specific target.
The specific target emotion classification device 8 based on attention coding and graph convolution network can include, but is not limited to, a processor 80 and a memory 81. Those skilled in the art will appreciate that FIG. 8 is merely an example of a specific target emotion classification device 8 based on attention coding and graph convolution network, and does not constitute a limitation on a specific target emotion classification device 8 based on attention coding and graph convolution network, and may include more or less components than those shown, or combine some components, or different components, for example, the specific target emotion classification device 8 based on attention coding and graph convolution network may further include an input-output device, a network access device, a bus, etc.
The Processor 80 may be a Central Processing Unit (CPU), other general purpose Processor, a Digital Signal Processor (DSP), an Application Specific Integrated Circuit (ASIC), a Field-Programmable Gate Array (FPGA) or other Programmable logic device, discrete Gate or transistor logic device, discrete hardware component, or the like. A general purpose processor may be a microprocessor or the processor may be any conventional processor or the like.
The storage 81 may be an internal storage unit of the attention coding and graph volume network based specific target emotion classification device 8, for example, a hard disk or a memory of the attention coding and graph volume network based specific target emotion classification device 8. The memory 81 may also be an external storage device of the specific target emotion classification device 8 based on the attention code and the graph volume network, such as a plug-in hard disk, a Smart Media Card (SMC), a Secure Digital (SD) Card, a Flash memory Card (Flash Card) and the like equipped on the specific target emotion classification device 8 based on the attention code and the graph volume network. Further, the memory 81 may also include both an internal storage unit and an external storage device of the specific target emotion classification device 8 based on the attention coding and the graph volume network. The memory 81 is used to store the computer program and other programs and data required by the specific target emotion classification apparatus based on attention coding and graph volume network. The memory 81 may also be used to temporarily store data that has been output or is to be output.
It will be apparent to those skilled in the art that, for convenience and brevity of description, only the above-mentioned division of the functional units and modules is illustrated, and in practical applications, the above-mentioned function distribution may be performed by different functional units and modules according to needs, that is, the internal structure of the apparatus is divided into different functional units or modules to perform all or part of the above-mentioned functions. Each functional unit and module in the embodiments may be integrated in one processing unit, or each unit may exist alone physically, or two or more units are integrated in one unit, and the integrated unit may be implemented in a form of hardware, or in a form of software functional unit. In addition, specific names of the functional units and modules are only for convenience of distinguishing from each other, and are not used for limiting the protection scope of the present application. The specific working processes of the units and modules in the system may refer to the corresponding processes in the foregoing method embodiments, and are not described herein again.
In the above embodiments, the descriptions of the respective embodiments have respective emphasis, and reference may be made to the related descriptions of other embodiments for parts that are not described or illustrated in a certain embodiment.
Those of ordinary skill in the art will appreciate that the various illustrative elements and algorithm steps described in connection with the embodiments disclosed herein may be implemented as electronic hardware or combinations of computer software and electronic hardware. Whether such functionality is implemented as hardware or software depends upon the particular application and design constraints imposed on the implementation. Skilled artisans may implement the described functionality in varying ways for each particular application, but such implementation decisions should not be interpreted as causing a departure from the scope of the present invention.
In the embodiments provided in the present invention, it should be understood that the disclosed apparatus/terminal device and method may be implemented in other ways. For example, the above-described embodiments of the apparatus/terminal device are merely illustrative, and for example, the division of the modules or units is only one logical division, and there may be other divisions when actually implemented, for example, a plurality of units or components may be combined or integrated into another system, or some features may be omitted, or not executed. In addition, the shown or discussed mutual coupling or direct coupling or communication connection may be an indirect coupling or communication connection through some interfaces, devices or units, and may be in an electrical, mechanical or other form.
The units described as separate parts may or may not be physically separate, and parts displayed as units may or may not be physical units, may be located in one place, or may be distributed on a plurality of network units. Some or all of the units can be selected according to actual needs to achieve the purpose of the solution of the embodiment.
In addition, functional units in the embodiments of the present invention may be integrated into one processing unit, or each unit may exist alone physically, or two or more units are integrated into one unit. The integrated unit can be realized in a form of hardware, and can also be realized in a form of a software functional unit.
The integrated modules/units, if implemented in the form of software functional units and sold or used as separate products, may be stored in a computer readable storage medium. Based on such understanding, all or part of the flow of the method according to the embodiments of the present invention may also be implemented by a computer program, which may be stored in a computer-readable storage medium, and when the computer program is executed by a processor, the steps of the method embodiments may be implemented. Wherein the computer program comprises computer program code, which may be in the form of source code, object code, an executable file or some intermediate form, etc. The computer-readable medium may include: any entity or device capable of carrying the computer program code, recording medium, usb disk, removable hard disk, magnetic disk, optical disk, computer Memory, Read-Only Memory (ROM), Random Access Memory (RAM), electrical carrier wave signals, telecommunications signals, software distribution medium, and the like. It should be noted that the computer readable medium may contain content that is subject to appropriate increase or decrease as required by legislation and patent practice in jurisdictions, for example, in some jurisdictions, computer readable media does not include electrical carrier signals and telecommunications signals as is required by legislation and patent practice. The present invention is not limited to the above-described embodiments, and various modifications and variations of the present invention are intended to be included within the scope of the claims and the equivalent technology of the present invention if they do not depart from the spirit and scope of the present invention.

Claims (10)

1. A specific target emotion classification method based on attention coding and graph convolution network is characterized by comprising the following steps:
obtaining a word vector corresponding to the context and a word vector corresponding to the specific target;
inputting the word vector corresponding to the context and the word vector corresponding to the specific target into a preset bidirectional recurrent neural network model to obtain a hidden state vector corresponding to the context and a hidden state vector corresponding to the specific target;
extracting a syntactic vector in a syntactic dependency tree corresponding to the context based on a preset graph convolutional neural network combined with point-to-point convolution;
performing multi-head self-attention coding on the hidden state vector corresponding to the context, the hidden state vector corresponding to the specific target and the syntax vector to respectively obtain context semantic information coding, specific target semantic information coding and syntax information coding;
respectively carrying out multi-head interactive attention coding on the context semantic information code and the syntax information code and the specific target semantic information code and the syntax information code to obtain a context-syntax information code and a specific target-syntax information code;
averaging and pooling the context semantic information codes, the context-syntax information codes and the specific target-syntax information codes, and then splicing to obtain a feature representation corresponding to the specific target;
and inputting the feature representation into a preset normalization index function to obtain an emotion classification result of the specific target.
2. The method for classifying emotion of specific target based on attention coding and graph convolution network as claimed in claim 1, wherein said method for inputting word vector corresponding to context and word vector corresponding to specific target into pre-defined bi-directional recurrent neural network model to obtain hidden state vector corresponding to context and hidden state vector corresponding to specific target comprises the steps of:
corresponding the context to a word vector
Figure FDA0002366196970000011
Word vectors corresponding to specific targets
Figure FDA0002366196970000012
Inputting the hidden state vector corresponding to the context and the hidden state vector corresponding to the specific target through a preset bidirectional cyclic neural network model LSTM, wherein the process comprises the following steps:
Figure FDA0002366196970000013
and
Figure FDA0002366196970000014
wherein n represents the dimension of the word vector corresponding to the context, m represents the dimension of the word vector corresponding to the specific target,
Figure FDA0002366196970000015
representing the forward operation process in the preset bidirectional recurrent neural network model LSTM,
Figure FDA0002366196970000016
representing the inverse operation process in a preset bidirectional recurrent neural network model LSTM, HcRepresenting a hidden state vector corresponding to the context, HtRepresenting the hidden state vector corresponding to the context.
3. The method for classifying specific target emotion based on attention coding and graph convolution network as claimed in claim 1, wherein the extracting the syntactic vector in the syntactic dependency tree corresponding to the context based on the preset graph convolution neural network combined with point-by-point convolution comprises the steps of:
according to the position and position weight distribution function of the specific target in the context, obtaining the position weight corresponding to each word in the context;
obtaining a syntactic dependency tree corresponding to a context;
obtaining an adjacency matrix corresponding to the words in the context according to the syntactic dependency tree, wherein the adjacency matrix reflects the adjacency relation of the words in the context;
inputting the adjacency matrix and the position weight corresponding to each word into a preset graph convolution neural network to obtain an output result of an output layer;
and performing point-by-point convolution on the output result to obtain the syntactic vector.
4. The method for classifying emotion of specific target based on attention coding and graph convolution network as claimed in claim 3, wherein said obtaining the position weight corresponding to each word in the context according to the position of specific target in the context and the position weight distribution function comprises the steps of:
according to the position and position weight distribution function of the specific target in the context, obtaining the position weight corresponding to each word in the context; wherein the location weight assignment function F (-) is as follows:
Figure FDA0002366196970000021
τ +1 represents the starting position of the specific target, m represents the number of words in the specific target, n represents the number of words in the context, q represents the number of words in the contextiIndicating the position weight of the ith word in the context.
5. The method for classifying specific target emotions based on attention coding and graph convolution network according to claim 4, wherein the step of inputting the adjacency matrix and the position weight of each word into a preset graph convolution neural network to obtain an output result of an output layer comprises the steps of:
inputting the adjacency matrix, the position weight of each word and the output result of the previous layer into a preset graph convolution operation formula to obtain the output result of the current layer, and repeatedly executing input operation until the output result of the output layer is obtained; the preset graph convolution operation formula is as follows:
Figure FDA0002366196970000022
Figure FDA0002366196970000031
Figure FDA0002366196970000032
Aijthe value representing the ith row and jth column of the adjacency matrix, A ∈ Rn×nRepresenting the contiguous matrix A as a matrix of n rows and n columns, AijRepresenting the output result of the previous layer of the convolutional neural network of the preset map, qlRepresenting the location weight of the jth word in the context,
Figure FDA0002366196970000033
the l-1 layer output result of the jth word of the convolutional neural network representing the preset graph,
Figure FDA0002366196970000034
l-level output result of j-th word representing preset graph convolutional neural network, diIndicates the depth, W, of the ith word in the syntactic dependency treelRepresenting the weight, bias blRepresenting a bias, RELU () representing an activation function.
6. The method for classifying specific target emotions based on attention coding and graph convolution network according to claim 3, wherein the point-by-point convolution is performed on the output result to obtain the syntax vector, and the method comprises the following steps:
inputting an output result of an output layer of a preset graph convolution network into a preset point-by-point convolution formula to obtain a syntax vector; wherein the preset point-by-point convolution formula is as follows:
Figure FDA0002366196970000035
PWC(h)=σ(h*Wpwc+bpwc)
hlrepresents the output result of the output layer of the preset graph convolution network,
Figure FDA0002366196970000036
represents a syntactic vector, where σ represents the activation function ReLu, a convolution operation,
Figure FDA0002366196970000037
is the learnable weight of the convolution kernel,
Figure FDA0002366196970000038
is the offset of the convolution kernel.
7. The method for classifying emotion of specific target based on attention coding and graph convolution network as claimed in claim 1, wherein multi-head self-attention coding is performed on hidden state vector corresponding to the context, hidden state vector corresponding to the specific target and syntax vector to obtain context semantic information code, specific target semantic information code and syntax information code, respectively, comprising the steps of:
the hidden state vector H corresponding to the contextcThe hidden state vector H corresponding to the specific targettAnd the syntactic vector
Figure FDA0002366196970000039
Respectively inputting a preset multi-head attention coding formula to obtain a context semantic information code HcsTarget-specific semantic information encoding HtsAnd syntax information coding Hgs(ii) a Wherein, the preset multi-head attention coding formula is as follows:
Hcs=MHA(Hc,Hc)
Hts=MHA(Ht,Ht)
Figure FDA0002366196970000041
Figure FDA0002366196970000042
oh=Attentionh(k,q)
Attention(k,q)=soft max(fs(k,q))k
Figure FDA0002366196970000043
fs(ki,qj)=tanh([ki;qj]·Watt)
fs(ki,qj) First input vector k ═ k, { representing multi-head attention1,k2,...,knAnd a second input vector q ═ q of multi-head attention1,q2,...,qmSemantic relevance of }, when multi-head self-attention coding is performed, k ≠ q, and when multi-head interactive attention coding is performed, k ≠ q, "; "refers to the concatenation of the vectors,
Figure FDA0002366196970000044
is a weight that can be learned that is,
Figure FDA0002366196970000045
represents 1 line 2dhidMatrix of columns, dhidRepresenting the dimension of the hidden state vector softmax () representing the normalized exponential function, Attentionh(k, q) and ohRepresenting the h-th output result in multi-head attention, h e [1, nhead],
Figure FDA0002366196970000046
Represents a pair ohD undergoing a linear changehidLine dhidA matrix of the columns is formed,
Figure FDA0002366196970000047
Figure FDA0002366196970000048
dhrepresenting the vector dimensions of the multi-headed attention-coding output.
8. The method for classifying specific target emotion based on attention coding and graph convolution network as claimed in claim 7, wherein said performing multi-head interactive attention coding on said context semantic information coding and syntax information coding, said specific target semantic information coding and syntax information coding respectively to obtain context-syntax information coding and specific target-syntax information coding, comprises the steps of:
encoding H the context semantic informationcsAnd syntax information coding HgsThe specific target semantic information code HtsAnd syntax information coding HgsRespectively inputting a preset multi-head attention coding formula to obtain a context-syntax information code HgtAnd specific target-syntax information encoding Hcg(ii) a Wherein the content of the first and second substances,
Hgt=MHA(Hgs,Hts)
Hcg=MHA(Hgs,Hcs)
Figure FDA0002366196970000051
9. the method for classifying emotion of specific target based on attention coding and graph convolution network as claimed in claim 1, wherein said method for obtaining feature representation corresponding to specific target by merging said context semantic information coding, said context-syntax information coding and said specific target-syntax information coding after averaging and pooling comprises the steps of:
encoding context-syntax information HcgSpecific object-syntax information coding HgtAnd context semantic coding HcsInputting a preset average pooling calculation formula, and splicing output results to obtain a feature expression u corresponding to the specific target; wherein, the preset average pooling calculation formula is as follows:
Figure FDA0002366196970000052
Figure FDA0002366196970000053
Figure FDA0002366196970000054
Figure FDA0002366196970000055
represents an average pooling result of context-syntax information encoding,
Figure FDA0002366196970000056
represents an average pooling result of a specific target-syntax information encoding,
Figure FDA0002366196970000057
representing average pooled results of context semantic coding, feature representation
Figure FDA0002366196970000058
"; "refers to the concatenation of vectors.
10. The method for classifying emotion of specific target based on attention coding and graph convolution network as claimed in claim 1, wherein said inputting said feature representation into a preset normalized exponential function to obtain emotion classification result of specific target comprises the steps of:
inputting the feature expression into a preset conversion formula, and then inputting the conversion result of the feature expression into a preset normalization index function to obtain the emotion classification result of the specific target, wherein the preset conversion formula and the preset normalization index function are as follows:
Figure FDA0002366196970000059
Figure FDA00023661969700000510
Figure FDA00023661969700000511
in order to be a representation of the feature,
Figure FDA00023661969700000512
and bu∈RcAre learnable weight and bias terms, respectively, y ∈ RcIs the emotion classification result of the specific object, and c represents the classified category.
CN202010036470.9A 2020-01-14 2020-01-14 Specific target emotion classification method based on attention coding and graph convolution network Active CN111259142B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202010036470.9A CN111259142B (en) 2020-01-14 2020-01-14 Specific target emotion classification method based on attention coding and graph convolution network

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202010036470.9A CN111259142B (en) 2020-01-14 2020-01-14 Specific target emotion classification method based on attention coding and graph convolution network

Publications (2)

Publication Number Publication Date
CN111259142A true CN111259142A (en) 2020-06-09
CN111259142B CN111259142B (en) 2020-12-25

Family

ID=70948730

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202010036470.9A Active CN111259142B (en) 2020-01-14 2020-01-14 Specific target emotion classification method based on attention coding and graph convolution network

Country Status (1)

Country Link
CN (1) CN111259142B (en)

Cited By (44)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111680159A (en) * 2020-06-11 2020-09-18 华东交通大学 Data processing method and device and electronic equipment
CN111710428A (en) * 2020-06-19 2020-09-25 华中师范大学 Biomedical text representation method for modeling global and local context interaction
CN111783474A (en) * 2020-07-16 2020-10-16 厦门市美亚柏科信息股份有限公司 Comment text viewpoint information processing method and device and storage medium
CN111881665A (en) * 2020-09-27 2020-11-03 华南师范大学 Word embedding representation method, device and equipment
CN112001187A (en) * 2020-08-26 2020-11-27 重庆理工大学 Emotion classification system based on Chinese syntax and graph convolution neural network
CN112001185A (en) * 2020-08-26 2020-11-27 重庆理工大学 Emotion classification method combining Chinese syntax and graph convolution neural network
CN112001186A (en) * 2020-08-26 2020-11-27 重庆理工大学 Emotion classification method using graph convolution neural network and Chinese syntax
CN112035661A (en) * 2020-08-24 2020-12-04 北京大学深圳研究生院 Text emotion analysis method and system based on graph convolution network and electronic device
CN112132223A (en) * 2020-09-27 2020-12-25 腾讯科技(深圳)有限公司 Graph pooling method, device, equipment and storage medium
CN112131383A (en) * 2020-08-26 2020-12-25 华南师范大学 Specific target emotion polarity classification method
CN112214601A (en) * 2020-10-21 2021-01-12 厦门市美亚柏科信息股份有限公司 Social short text sentiment classification method and device and storage medium
CN112417157A (en) * 2020-12-15 2021-02-26 华南师范大学 Emotion classification method of text attribute words based on deep learning network
CN112528672A (en) * 2020-12-14 2021-03-19 北京邮电大学 Aspect-level emotion analysis method and device based on graph convolution neural network
CN112579778A (en) * 2020-12-23 2021-03-30 重庆邮电大学 Aspect-level emotion classification method based on multi-level feature attention
CN112633010A (en) * 2020-12-29 2021-04-09 山东师范大学 Multi-head attention and graph convolution network-based aspect-level emotion analysis method and system
CN112686034A (en) * 2021-03-22 2021-04-20 华南师范大学 Emotion classification method, device and equipment
CN112686056A (en) * 2021-03-22 2021-04-20 华南师范大学 Emotion classification method
CN112733764A (en) * 2021-01-15 2021-04-30 天津大学 Method for recognizing video emotion information based on multiple modes
CN112801219A (en) * 2021-03-22 2021-05-14 华南师范大学 Multi-mode emotion classification method, device and equipment
CN112818670A (en) * 2020-08-05 2021-05-18 百度(美国)有限责任公司 Split syntax and semantics in a decomposable variational auto-encoder sentence representation
CN112860907A (en) * 2021-04-27 2021-05-28 华南师范大学 Emotion classification method and equipment
CN112883741A (en) * 2021-04-29 2021-06-01 华南师范大学 Specific target emotion classification method based on dual-channel graph neural network
CN112926337A (en) * 2021-02-05 2021-06-08 昆明理工大学 End-to-end aspect level emotion analysis method combined with reconstructed syntax information
CN112966074A (en) * 2021-05-17 2021-06-15 华南师范大学 Emotion analysis method and device, electronic equipment and storage medium
CN113033215A (en) * 2021-05-18 2021-06-25 华南师范大学 Emotion detection method, device, equipment and storage medium
CN113177562A (en) * 2021-04-29 2021-07-27 京东数字科技控股股份有限公司 Vector determination method and device based on self-attention mechanism fusion context information
CN113220887A (en) * 2021-05-31 2021-08-06 华南师范大学 Emotion classification method and device using target knowledge enhancement model
CN113326739A (en) * 2021-05-07 2021-08-31 山东大学 Online learning participation degree evaluation method based on space-time attention network, evaluation system, equipment and storage medium
CN113468874A (en) * 2021-06-09 2021-10-01 大连理工大学 Biomedical relation extraction method based on graph convolution self-coding
CN113505583A (en) * 2021-05-27 2021-10-15 山东交通学院 Sentiment reason clause pair extraction method based on semantic decision diagram neural network
CN113535904A (en) * 2021-07-23 2021-10-22 重庆邮电大学 Aspect level emotion analysis method based on graph neural network
CN113609267A (en) * 2021-07-21 2021-11-05 上海交通大学 Speech relation recognition method and system based on GCNDT-MacBERT neural network framework
CN113688212A (en) * 2021-10-27 2021-11-23 华南师范大学 Sentence emotion analysis method, device and equipment
US11194972B1 (en) 2021-02-19 2021-12-07 Institute Of Automation, Chinese Academy Of Sciences Semantic sentiment analysis method fusing in-depth features and time sequence models
WO2022063151A1 (en) * 2020-09-23 2022-03-31 Beijing Wodong Tianjun Information Technology Co., Ltd. Method and system for relation learning by multi-hop attention graph neural network
CN114386389A (en) * 2020-10-22 2022-04-22 四川大学 Aspect emotion analysis method based on joint learning
CN114579707A (en) * 2022-03-07 2022-06-03 桂林旅游学院 BERT neural network and multi-semantic learning-based aspect-level emotion analysis method
CN114677234A (en) * 2022-04-26 2022-06-28 河南大学 Graph convolution neural network social contact recommendation method and system integrating multi-channel attention mechanism
CN114707518A (en) * 2022-06-08 2022-07-05 四川大学 Semantic fragment-oriented target emotion analysis method, device, equipment and medium
CN114881042A (en) * 2022-06-02 2022-08-09 电子科技大学 Chinese emotion analysis method based on graph convolution network fusion syntax dependence and part of speech
CN115860006A (en) * 2023-02-13 2023-03-28 广东工业大学 Aspect level emotion prediction method and device based on semantic syntax
WO2023050470A1 (en) * 2021-09-30 2023-04-06 深圳市检验检疫科学研究院 Event detection method and apparatus based on multi-layer graph attention network
WO2023077562A1 (en) * 2021-11-03 2023-05-11 深圳市检验检疫科学研究院 Graph perturbation strategy-based event detection method and apparatus
CN117592460A (en) * 2024-01-17 2024-02-23 西安邮电大学 End-to-end face level emotion analysis method based on feature fusion

Citations (11)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20160132789A1 (en) * 2013-09-30 2016-05-12 Manyworlds, Inc. Streams of Attention Method, System, and Apparatus
CN108399158A (en) * 2018-02-05 2018-08-14 华南理工大学 Attribute sensibility classification method based on dependency tree and attention mechanism
US20180261214A1 (en) * 2017-02-06 2018-09-13 Facebook, Inc. Sequence-to-sequence convolutional architecture
CN108549646A (en) * 2018-04-24 2018-09-18 中译语通科技股份有限公司 A kind of neural network machine translation system based on capsule, information data processing terminal
CN108664632A (en) * 2018-05-15 2018-10-16 华南理工大学 A kind of text emotion sorting algorithm based on convolutional neural networks and attention mechanism
CN109408823A (en) * 2018-10-31 2019-03-01 华南师范大学 A kind of specific objective sentiment analysis method based on multi-channel model
CN109446331A (en) * 2018-12-07 2019-03-08 华中科技大学 A kind of text mood disaggregated model method for building up and text mood classification method
CN109558487A (en) * 2018-11-06 2019-04-02 华南师范大学 Document Classification Method based on the more attention networks of hierarchy
CN109582764A (en) * 2018-11-09 2019-04-05 华南师范大学 Interaction attention sentiment analysis method based on interdependent syntax
CN109657246A (en) * 2018-12-19 2019-04-19 中山大学 A kind of extraction-type machine reading based on deep learning understands the method for building up of model
CN109885670A (en) * 2019-02-13 2019-06-14 北京航空航天大学 A kind of interaction attention coding sentiment analysis method towards topic text

Patent Citations (11)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20160132789A1 (en) * 2013-09-30 2016-05-12 Manyworlds, Inc. Streams of Attention Method, System, and Apparatus
US20180261214A1 (en) * 2017-02-06 2018-09-13 Facebook, Inc. Sequence-to-sequence convolutional architecture
CN108399158A (en) * 2018-02-05 2018-08-14 华南理工大学 Attribute sensibility classification method based on dependency tree and attention mechanism
CN108549646A (en) * 2018-04-24 2018-09-18 中译语通科技股份有限公司 A kind of neural network machine translation system based on capsule, information data processing terminal
CN108664632A (en) * 2018-05-15 2018-10-16 华南理工大学 A kind of text emotion sorting algorithm based on convolutional neural networks and attention mechanism
CN109408823A (en) * 2018-10-31 2019-03-01 华南师范大学 A kind of specific objective sentiment analysis method based on multi-channel model
CN109558487A (en) * 2018-11-06 2019-04-02 华南师范大学 Document Classification Method based on the more attention networks of hierarchy
CN109582764A (en) * 2018-11-09 2019-04-05 华南师范大学 Interaction attention sentiment analysis method based on interdependent syntax
CN109446331A (en) * 2018-12-07 2019-03-08 华中科技大学 A kind of text mood disaggregated model method for building up and text mood classification method
CN109657246A (en) * 2018-12-19 2019-04-19 中山大学 A kind of extraction-type machine reading based on deep learning understands the method for building up of model
CN109885670A (en) * 2019-02-13 2019-06-14 北京航空航天大学 A kind of interaction attention coding sentiment analysis method towards topic text

Non-Patent Citations (2)

* Cited by examiner, † Cited by third party
Title
ZHAOA, PINLONG等: "Modeling Sentiment Dependencies with Graph Convolutional Networks for Aspect-level Sentiment Classification", 《HTTP://AU.ARXIV.ORG/》 *
梁丕军: "基于循环神经网络和卷积神经网络的中文情感分类研究", 《中国优秀硕士学位论文全文数据库》 *

Cited By (71)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111680159A (en) * 2020-06-11 2020-09-18 华东交通大学 Data processing method and device and electronic equipment
CN111680159B (en) * 2020-06-11 2023-08-29 华东交通大学 Data processing method and device and electronic equipment
CN111710428B (en) * 2020-06-19 2022-05-31 华中师范大学 Biomedical text representation method for modeling global and local context interaction
CN111710428A (en) * 2020-06-19 2020-09-25 华中师范大学 Biomedical text representation method for modeling global and local context interaction
CN111783474A (en) * 2020-07-16 2020-10-16 厦门市美亚柏科信息股份有限公司 Comment text viewpoint information processing method and device and storage medium
CN111783474B (en) * 2020-07-16 2023-04-07 厦门市美亚柏科信息股份有限公司 Comment text viewpoint information processing method and device and storage medium
CN112818670B (en) * 2020-08-05 2023-10-27 百度(美国)有限责任公司 Segmentation grammar and semantics in a decomposable variant automatic encoder sentence representation
CN112818670A (en) * 2020-08-05 2021-05-18 百度(美国)有限责任公司 Split syntax and semantics in a decomposable variational auto-encoder sentence representation
CN112035661A (en) * 2020-08-24 2020-12-04 北京大学深圳研究生院 Text emotion analysis method and system based on graph convolution network and electronic device
CN112001185A (en) * 2020-08-26 2020-11-27 重庆理工大学 Emotion classification method combining Chinese syntax and graph convolution neural network
CN112131383A (en) * 2020-08-26 2020-12-25 华南师范大学 Specific target emotion polarity classification method
CN112001186A (en) * 2020-08-26 2020-11-27 重庆理工大学 Emotion classification method using graph convolution neural network and Chinese syntax
CN112001187A (en) * 2020-08-26 2020-11-27 重庆理工大学 Emotion classification system based on Chinese syntax and graph convolution neural network
CN112001185B (en) * 2020-08-26 2021-07-20 重庆理工大学 Emotion classification method combining Chinese syntax and graph convolution neural network
CN112001187B (en) * 2020-08-26 2021-05-28 重庆理工大学 Emotion classification system based on Chinese syntax and graph convolution neural network
WO2022063151A1 (en) * 2020-09-23 2022-03-31 Beijing Wodong Tianjun Information Technology Co., Ltd. Method and system for relation learning by multi-hop attention graph neural network
CN111881665B (en) * 2020-09-27 2021-01-05 华南师范大学 Word embedding representation method, device and equipment
CN112132223A (en) * 2020-09-27 2020-12-25 腾讯科技(深圳)有限公司 Graph pooling method, device, equipment and storage medium
CN111881665A (en) * 2020-09-27 2020-11-03 华南师范大学 Word embedding representation method, device and equipment
CN112132223B (en) * 2020-09-27 2024-02-27 腾讯科技(深圳)有限公司 Method, device, equipment and storage medium for pooling
CN112214601A (en) * 2020-10-21 2021-01-12 厦门市美亚柏科信息股份有限公司 Social short text sentiment classification method and device and storage medium
CN112214601B (en) * 2020-10-21 2022-06-10 厦门市美亚柏科信息股份有限公司 Social short text sentiment classification method and device and storage medium
CN114386389A (en) * 2020-10-22 2022-04-22 四川大学 Aspect emotion analysis method based on joint learning
CN114386389B (en) * 2020-10-22 2023-06-06 四川大学 Aspect emotion analysis method based on joint learning
CN112528672B (en) * 2020-12-14 2021-07-30 北京邮电大学 Aspect-level emotion analysis method and device based on graph convolution neural network
CN112528672A (en) * 2020-12-14 2021-03-19 北京邮电大学 Aspect-level emotion analysis method and device based on graph convolution neural network
CN112417157A (en) * 2020-12-15 2021-02-26 华南师范大学 Emotion classification method of text attribute words based on deep learning network
CN112579778A (en) * 2020-12-23 2021-03-30 重庆邮电大学 Aspect-level emotion classification method based on multi-level feature attention
CN112633010B (en) * 2020-12-29 2023-08-04 山东师范大学 Aspect-level emotion analysis method and system based on multi-head attention and graph convolution network
CN112633010A (en) * 2020-12-29 2021-04-09 山东师范大学 Multi-head attention and graph convolution network-based aspect-level emotion analysis method and system
CN112733764A (en) * 2021-01-15 2021-04-30 天津大学 Method for recognizing video emotion information based on multiple modes
CN112926337A (en) * 2021-02-05 2021-06-08 昆明理工大学 End-to-end aspect level emotion analysis method combined with reconstructed syntax information
US11194972B1 (en) 2021-02-19 2021-12-07 Institute Of Automation, Chinese Academy Of Sciences Semantic sentiment analysis method fusing in-depth features and time sequence models
CN112801219A (en) * 2021-03-22 2021-05-14 华南师范大学 Multi-mode emotion classification method, device and equipment
CN112686056A (en) * 2021-03-22 2021-04-20 华南师范大学 Emotion classification method
CN112801219B (en) * 2021-03-22 2021-06-18 华南师范大学 Multi-mode emotion classification method, device and equipment
CN112686034A (en) * 2021-03-22 2021-04-20 华南师范大学 Emotion classification method, device and equipment
CN112860907B (en) * 2021-04-27 2021-06-29 华南师范大学 Emotion classification method and equipment
CN112860907A (en) * 2021-04-27 2021-05-28 华南师范大学 Emotion classification method and equipment
CN113177562A (en) * 2021-04-29 2021-07-27 京东数字科技控股股份有限公司 Vector determination method and device based on self-attention mechanism fusion context information
CN113177562B (en) * 2021-04-29 2024-02-06 京东科技控股股份有限公司 Vector determination method and device for merging context information based on self-attention mechanism
CN112883741A (en) * 2021-04-29 2021-06-01 华南师范大学 Specific target emotion classification method based on dual-channel graph neural network
CN113326739A (en) * 2021-05-07 2021-08-31 山东大学 Online learning participation degree evaluation method based on space-time attention network, evaluation system, equipment and storage medium
CN113326739B (en) * 2021-05-07 2022-08-09 山东大学 Online learning participation degree evaluation method based on space-time attention network, evaluation system, equipment and storage medium
CN112966074A (en) * 2021-05-17 2021-06-15 华南师范大学 Emotion analysis method and device, electronic equipment and storage medium
CN113033215B (en) * 2021-05-18 2021-08-13 华南师范大学 Emotion detection method, device, equipment and storage medium
CN113033215A (en) * 2021-05-18 2021-06-25 华南师范大学 Emotion detection method, device, equipment and storage medium
CN113505583B (en) * 2021-05-27 2023-07-18 山东交通学院 Emotion reason clause pair extraction method based on semantic decision graph neural network
CN113505583A (en) * 2021-05-27 2021-10-15 山东交通学院 Sentiment reason clause pair extraction method based on semantic decision diagram neural network
CN113220887A (en) * 2021-05-31 2021-08-06 华南师范大学 Emotion classification method and device using target knowledge enhancement model
CN113468874A (en) * 2021-06-09 2021-10-01 大连理工大学 Biomedical relation extraction method based on graph convolution self-coding
CN113468874B (en) * 2021-06-09 2024-04-16 大连理工大学 Biomedical relation extraction method based on graph convolution self-coding
CN113609267B (en) * 2021-07-21 2023-11-07 上海交通大学 Speech relation recognition method and system based on GCNDT-MacBERT neural network framework
CN113609267A (en) * 2021-07-21 2021-11-05 上海交通大学 Speech relation recognition method and system based on GCNDT-MacBERT neural network framework
CN113535904A (en) * 2021-07-23 2021-10-22 重庆邮电大学 Aspect level emotion analysis method based on graph neural network
CN113535904B (en) * 2021-07-23 2022-08-09 重庆邮电大学 Aspect level emotion analysis method based on graph neural network
WO2023050470A1 (en) * 2021-09-30 2023-04-06 深圳市检验检疫科学研究院 Event detection method and apparatus based on multi-layer graph attention network
CN113688212A (en) * 2021-10-27 2021-11-23 华南师范大学 Sentence emotion analysis method, device and equipment
CN113688212B (en) * 2021-10-27 2022-02-08 华南师范大学 Sentence emotion analysis method, device and equipment
WO2023077562A1 (en) * 2021-11-03 2023-05-11 深圳市检验检疫科学研究院 Graph perturbation strategy-based event detection method and apparatus
CN114579707A (en) * 2022-03-07 2022-06-03 桂林旅游学院 BERT neural network and multi-semantic learning-based aspect-level emotion analysis method
CN114579707B (en) * 2022-03-07 2023-07-28 桂林旅游学院 Aspect-level emotion analysis method based on BERT neural network and multi-semantic learning
CN114677234A (en) * 2022-04-26 2022-06-28 河南大学 Graph convolution neural network social contact recommendation method and system integrating multi-channel attention mechanism
CN114677234B (en) * 2022-04-26 2024-04-30 河南大学 Graph convolution neural network social recommendation method and system integrating multichannel attention mechanisms
CN114881042B (en) * 2022-06-02 2023-05-02 电子科技大学 Chinese emotion analysis method based on graph-convolution network fusion of syntactic dependency and part of speech
CN114881042A (en) * 2022-06-02 2022-08-09 电子科技大学 Chinese emotion analysis method based on graph convolution network fusion syntax dependence and part of speech
CN114707518B (en) * 2022-06-08 2022-08-16 四川大学 Semantic fragment-oriented target emotion analysis method, device, equipment and medium
CN114707518A (en) * 2022-06-08 2022-07-05 四川大学 Semantic fragment-oriented target emotion analysis method, device, equipment and medium
CN115860006B (en) * 2023-02-13 2023-04-25 广东工业大学 Aspect-level emotion prediction method and device based on semantic syntax
CN115860006A (en) * 2023-02-13 2023-03-28 广东工业大学 Aspect level emotion prediction method and device based on semantic syntax
CN117592460A (en) * 2024-01-17 2024-02-23 西安邮电大学 End-to-end face level emotion analysis method based on feature fusion

Also Published As

Publication number Publication date
CN111259142B (en) 2020-12-25

Similar Documents

Publication Publication Date Title
CN111259142B (en) Specific target emotion classification method based on attention coding and graph convolution network
CN112131383B (en) Specific target emotion polarity classification method
CN112966074B (en) Emotion analysis method and device, electronic equipment and storage medium
CN107516110B (en) Medical question-answer semantic clustering method based on integrated convolutional coding
CN112417157B (en) Emotion classification method of text attribute words based on deep learning network
CN111476023B (en) Method and device for identifying entity relationship
CN111444344B (en) Entity classification method, entity classification device, computer equipment and storage medium
CN111738001B (en) Training method of synonym recognition model, synonym determination method and equipment
CN114565104A (en) Language model pre-training method, result recommendation method and related device
CN111191457B (en) Natural language semantic recognition method, device, computer equipment and storage medium
CN111078847A (en) Power consumer intention identification method and device, computer equipment and storage medium
CN111881671B (en) Attribute word extraction method
CN112163099A (en) Text recognition method and device based on knowledge graph, storage medium and server
CN112749274B (en) Chinese text classification method based on attention mechanism and interference word deletion
CN113722474A (en) Text classification method, device, equipment and storage medium
CN114676704A (en) Sentence emotion analysis method, device and equipment and storage medium
CN116663568B (en) Critical task identification system and method based on priority
CN111428513A (en) False comment analysis method based on convolutional neural network
CN116010586A (en) Method, device, equipment and storage medium for generating health advice
CN114528835A (en) Semi-supervised specialized term extraction method, medium and equipment based on interval discrimination
CN114331122A (en) Key person risk level assessment method and related equipment
CN112749737A (en) Image classification method and device, electronic equipment and storage medium
CN113239702A (en) Intention recognition method and device and electronic equipment
CN116304020A (en) Industrial text entity extraction method based on semantic source analysis and span characteristics
CN117314593B (en) Insurance item pushing method and system based on user behavior analysis

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
CB03 Change of inventor or designer information
CB03 Change of inventor or designer information

Inventor after: Xiao Luwei

Inventor after: Hu Xiaohui

Inventor after: Liu Yifan

Inventor after: Xue Yun

Inventor after: Gu Donghong

Inventor after: Chen Bingliang

Inventor before: Xiao Luwei

Inventor before: Liu Yifan

Inventor before: Hu Xiaohui

Inventor before: Xue Yun

Inventor before: Gu Donghong

Inventor before: Chen Bingliang

GR01 Patent grant
GR01 Patent grant
TR01 Transfer of patent right
TR01 Transfer of patent right

Effective date of registration: 20210708

Address after: 210012 4th floor, building C, Wanbo Science Park, 20 Fengxin Road, Yuhuatai District, Nanjing City, Jiangsu Province

Patentee after: NANJING SILICON INTELLIGENCE TECHNOLOGY Co.,Ltd.

Address before: Room 614-615, No.1, Lane 2277, Zuchongzhi Road, China (Shanghai) pilot Free Trade Zone, Pudong New Area, Shanghai, 200120

Patentee before: Shanghai Airlines Intellectual Property Services Ltd.

Effective date of registration: 20210708

Address after: Room 614-615, No.1, Lane 2277, Zuchongzhi Road, China (Shanghai) pilot Free Trade Zone, Pudong New Area, Shanghai, 200120

Patentee after: Shanghai Airlines Intellectual Property Services Ltd.

Address before: School of physics and telecommunication engineering, South China Normal University, No. 378, Waihuan West Road, Panyu District, Guangzhou City, Guangdong Province, 510006

Patentee before: SOUTH CHINA NORMAL University