CN110852070A - Document vector generation method - Google Patents

Document vector generation method Download PDF

Info

Publication number
CN110852070A
CN110852070A CN201911025383.7A CN201911025383A CN110852070A CN 110852070 A CN110852070 A CN 110852070A CN 201911025383 A CN201911025383 A CN 201911025383A CN 110852070 A CN110852070 A CN 110852070A
Authority
CN
China
Prior art keywords
document
word
attention
list
words
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN201911025383.7A
Other languages
Chinese (zh)
Inventor
金霞
杨红飞
张庭正
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Hangzhou Firestone Technology Co Ltd
Original Assignee
Hangzhou Firestone Technology Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Hangzhou Firestone Technology Co Ltd filed Critical Hangzhou Firestone Technology Co Ltd
Priority to CN201911025383.7A priority Critical patent/CN110852070A/en
Publication of CN110852070A publication Critical patent/CN110852070A/en
Pending legal-status Critical Current

Links

Images

Landscapes

  • Machine Translation (AREA)

Abstract

The invention discloses a document vector generating method, which comprises the steps of firstly converting words of each sentence in a document into a list, then forming the list of the document by the list of each sentence, finally mapping each word in the document list to a unique integer number, and then sliding a window in the document list to obtain a training sample of a hierarchical attention network; secondly, constructing a hierarchical attention network comprising a word-level encoder, a word-level attention layer, a sentence-level encoder, a sentence-level attention layer and document attention; and then constructing a language model framework, wherein the hierarchical attention network and the language model framework form a language model, training the language model, fitting parameters to be trained in the model, and inputting the prediction data into the model after the language model is trained to obtain a document vector of the model. When the method is used, partial variables do not need to be trained again for each different document, and the use cost is greatly reduced.

Description

Document vector generation method
Technical Field
The invention relates to the field of document vector generation, in particular to a document vector generation method.
Background
In recent years, the use of deep learning methods to learn low-dimensional real-valued vector expressions for words, sentences or documents has been widely applied in the field of natural language processing, and in some natural language processing tasks of text classification, text similarity calculation and emotion analysis, the quality of a document vector directly affects the result of the task.
The existing method for generating the document feature vector is based on a word2vec (Le Q V, MikolovT. distributed retrieval of sequences and Documents [ J ].2014.) word vector model, a document feature vector is added in the training process of a language model, and the document vector is directly trained while the word vector is trained. Although the document vector model captures semantic and sequence information among words, the range capable of being captured is limited; and because the document vectors are directly trained, the vectors of different documents need to be retrained, so that the document vectors of the training set need to be trained instead of being directly predicted.
The existing document vector generation method is generally difficult to pay attention to the whole document information, and when a trained model is used for predicting a vector of a new document, partial variables need to be retrained, which is high in cost in a production environment.
Disclosure of Invention
The invention aims to provide a document vector generation method based on a hierarchical attention network, aiming at overcoming the defects of the prior art, and the method can capture the information of the whole document by using the attention network and does not need to retrain any variable during prediction.
The purpose of the invention is realized by the following technical scheme: a document vector generation method, the method comprising the steps of:
(1) preparing data;
(1.1) converting words of each sentence in the document into a list, forming the list of the document by the list of each sentence, and finally mapping each word in the document list to a unique integer number which is marked as doc;
(1.2) sliding a window with the window size of window _ size in a document list, recording a word list in the window as words, and recording the second window _ size//2 words in the window as label; one document will eventually form a plurality of words and labels, while the data of one sample will include doc, words and label, and one document will eventually form a plurality of samples as training samples of the hierarchical attention network, which samples have the same doc, but different words and labels.
(2) Constructing a hierarchical attention network; the hierarchical attention network comprises a word-level encoder word encoder, a word-level attention layer word attention, a sentence-level encoder sensor encoder, a sentence-level attention layer sensor attention and a document attention doc attention; the word encoder layer and the sensor encoder layer are both of a sequence encoder layer type; the word attribute layer and the presence attribute layer are attention networks, the context text context of the word attribute layer and the presence attribute layer are different, and the context of the word attribute layer is uwAnd context of sentenceatention layer is us
(3) Language model framework: splicing word vectors corresponding to doc _ attribute and words output by the hierarchical attention network, and mapping to label by adopting softmax; the word vector is an embedded matrix WeBy the integer number of words from the embedded matrix WeIndexing to the corresponding word vector;
(4) the hierarchical attention network and the language model framework form a language model, the language model is trained, parameters to be trained in the model are fitted, and doc _ attention corresponding to each document is a final document vector after the language model is trained; the prediction data is input into the model, and a document vector of the prediction data can be obtained.
Further, in step (1.1), words of each sentence in the document may be converted into a list, the list of each sentence constitutes the list of the document, and finally, each word in the list of the document is mapped to a unique integer number.
Further, in step (1.2), both words and label are the result of the words mapped to integer numbers.
Further, in step (1.2), the window size window _ size value is 8.
Further, in step (1.2), window _ size//2 represents the integer part of the window size divided by 2.
Further, in step (2), the sequence encoder is based on a GRU (gated current unit) network, the GRU updates the hidden layer state, mainly through an association gate and an update gate, and keeps the hidden state at current step t as htThen the update gate is:
Figure RE-GDA0002318502020000021
wherein the content of the first and second substances,
Γu=σ(Wu[ht-1,xt]T+bu)
where σ is the sigmoid function, WuAnd buIs a parameter to be trained and is,
Figure RE-GDA0002318502020000022
to associate the door, xtIs the input of the GRU;
the associated gates are as follows:
Figure RE-GDA0002318502020000023
wherein tanh represents a hyperbolic tangent function, WCAnd bCIs a parameter to be trained;
the associated gate weights are as follows:
Γr=σ(Wr[ht-1,xt]T+br)
in the formula, WrAnd brIs the parameter to be trained.
Further, in the step (2), the word-level encoder word encoder is specifically:
xit=Wewit,t∈[1,T]
Figure RE-GDA0002318502020000031
Figure RE-GDA0002318502020000032
in the formula, xitFor the result of vectorization of the integer number corresponding to a word, WeIs an embedded matrix, witIs the vector after the integer number one-hot coding OneHot in words, T is the maximum value of the word number of each sentence,
Figure RE-GDA0002318502020000033
and
Figure RE-GDA0002318502020000034
the outputs of the GRUs in forward and reverse directions, respectively.
Further, in the step (2), the syntax-level encoder transmit encoder is specifically:
Figure RE-GDA0002318502020000035
Figure RE-GDA0002318502020000036
in the formula, siIs the presence _ entry of the ith sentence, and L is the number of sentences.
Further, in the step (2), the word-level attention layer word attention specifically includes:
uit=tanh(Wwhit+bw)
Figure RE-GDA0002318502020000037
Figure RE-GDA0002318502020000038
in the formula, hitBy passingAnd
Figure RE-GDA00023185020200000310
is spliced to obtain siIs the sense _ entry, W, of the ith sentencewAnd bwIs the parameter to be trained.
Further, in step (2), the sentence-level attention layer presence _ entry specifically includes:
ui=tanh(Wshi+bs)
Figure RE-GDA00023185020200000311
Figure RE-GDA00023185020200000312
in the formula, hiBy passingAnd
Figure RE-GDA00023185020200000314
is spliced to obtainsAnd bsIs the parameter to be trained.
The invention has the beneficial effects that: the present invention uses a hierarchical attention network to obtain document vectors: and mapping words in the clause to the content entry, and then mapping the information of the content entry to the final document vector by using the information of the content entry, so that the document vector is ensured to contain the information of the whole document, and the finally generated document vector does not lose semantic information among the words depending on a language model. Because the document vector is generated by using the hierarchical attention network, the hidden layer weight required by the generation of the final document vector is fitted in the model training process, and when the model is used, partial variables do not need to be trained again for each different document, so that the use cost is greatly reduced.
Drawings
FIG. 1 is a diagram of a hierarchical attention network framework of the present invention;
FIG. 2 is a diagram of a language model framework according to the present invention.
Detailed Description
The following describes embodiments of the present invention in further detail with reference to the accompanying drawings.
As shown in FIG. 1, the document vector generation method provided by the present invention is based on a language model, and embeds documents into the language model in the form of a hierarchical attention network. The method comprises the following specific steps:
(1) preparing data: for each document, the following processing is performed:
1) converting the words of each sentence in the document into a list, forming the list of the document by the list of each sentence, and finally mapping each word in the document list to a unique integer number and marking as doc.
For example, the original document is as follows:
the industry universal recommendation system includes two stages of flow, matching (match) and ranking (rank). In the matching process, a potential commodity set which is possibly interested is found mainly according to some interest points of the user. Because the massive nature of the whole commodity set, it is impractical for a user accessing in real time to calculate the interest level of the user in all commodities, so that a potential commodity set which the user may be interested in needs to be searched in advance according to some interests, feature strategies and the like, on the basis, deep ranking of interest categories of the commodity set is performed according to a specific model algorithm, and effect indexes are often quantified through Click rate (Click TroughRate), conversion rate, duration and the like, so that the main purpose of the rank stage is to predict the CTR and the like of a user in the commodity which the user is interested in, perform ranking according to the size of the predicted score, and finally return the ranking as the recommendation result of the recommendation system.
The list of converted documents is as follows:
[ [ 'industrial', 'etc', 'industrial',
'in', 'matching', 'in', 'course', 'in', 'main', 'to', 'is', 'root', 'in', 'use', 'user', 'of', 'one', 'some', 'with', 'interest', 'point', 'place', 'to', 'can', 'sense', 'like', 'can', 'latent', 'is', 'in', 'quotient', 'product', 'set',
the 'is a little bit of the' is a 'from', 'is a', 'is', the 'can', the ' g ', ' h ', ' R ', ' a ', ' T ', ' e ', ' a ', ' when ', ' long ', ' etc ', ' in ', ' amount ', ' a ','d ', ' in ', ' a ', ' n ', ' k ', ' a ', ' stage ', ' segment ', ' a ', ' main ', ' an ' of ', ' eye ', ' just ', ' in ', ' pre ', ' measuring ', ' a ', ' in ', ' a ', ' using ', ' a ' of the, the term "pre '," measure', 'divide', 'value', 'large', 'small', 'enter', 'line', 'arrange', 'sequence', 'most', 'final', 'back', 'arrange', 'sequence', 'do', 'is', 'push', 'recommend', 'system', 'is', 'push', 'recommend', 'result' ], or the like
After mapping the document list (as long as it is ensured that each literal can be mapped to a unique integer number) to an integer number:
[[599,227,773,153,138,537,1549,248,1533,693,4371,84,39,3104,465,2,954,1085,1, 1366,435,18,529,393,801,1032,972,17,6,864,2756,18,1913,393,1435,1575,17],
[5,1366,435,2,191,1085,35,449,80,12,653,121,138,385,2,29,1201,624,674,86,1, 616,56,88,90,736,624,674,2,2126,5,302,433,422],
[82,62,985,39,302,433,422,2,363,218,311,1,32,504,96,1841,730,2,138,385,164, 928,818,130,32,186,268,302,433,2,736,624,674,1085,250,12,42,504,1210,2,1,146,46, 133,80,546,336,653,121,29,1201,624,674,4,479,2037,1368,735,22,109,1668,616,138, 385,88,90,736,624,674,2,2126,5,302,433,422,1,5,327,367,4594,38,1,147,653,121,479, 499,2,1068,232,818,370,109,590,261,302,433,422,624,674,262,242,315,864,2756,1, 1103,1171,166,535,778,778,153,191,86,3210,234,18,140,1153,717,1032,1575,0,332, 1913,2034,2063,633,972,0,510,393,801,785,17,4,313,220,234,4,96,241,22,109,218, 220,1,146,46,0,1913,393,1435,1575,0,3104,465,2,449,80,915,2,55,5,62,546,769,29, 39,138,385,5,130,736,624,674,2,302,433,35,2,0,140,332,510,0,22,1,53,169,653,121, 546,769,262,270,2,50,144,590,261,864,2756,1,131,1200,1459,407,864,2756,580,15, 537,1549,248,1533,2,537,1549,1223,1171]]
in this step, each word in the document may also be converted into a list, and then the list of each word constitutes the list of the document, and finally each word in the document list is mapped to a unique integer number.
2) A window with a size of window size (window size 8 is used in this example) is slid over the document, the list of words in the window is denoted words, the first window size//2 (integer part of the quotient of window size divided by 2) words in the window are denoted label, where words and label are also the result of the words mapped to integer numbers. A document will eventually have multiple words and labels, while the data for a sample will include doc, words and label, and a document will eventually form multiple samples as training samples for the hierarchical attention network, with doc being the same but word and label being different.
(2) Hierarchical attention network: because two layers of attention networks are used, it is called a hierarchical attention network. Each sentence of doc passes through a word encoder and a word attention layer in sequence to obtain a sensor _ attention layer, and the sensor _ attention layer of each sentence passes through a sensor encoder and a sensor attention layer in sequence to obtain a final doc _ attention layer; both the word encoder layer and the sensor encoder layer are of a sequence encoder layer type; both the word attribute layer and the presence attribute layer are attention networks, except that their context text context is different.
The sequence encoder is based on a GRU (gated Current Unit) network, the GRU updates hidden state mainly through an association gate and an update gate, the GRU network can be replaced by other Recurrent neural networks, such as LSTM (Long Short Term memory), and the hidden state of the current step t is recorded as htThen the update gate is:
Figure RE-GDA0002318502020000061
wherein the content of the first and second substances,
Γu=σ(Wu[ht-1,xt]T+bu)
where σ is the sigmoid function, WuAnd buIs a parameter to be trained and is,
Figure RE-GDA0002318502020000062
to associate the door, xtIs the input of the GRU;
the associated gates are as follows:
Figure RE-GDA0002318502020000063
wherein tanh represents a hyperbolic tangent function, WCAnd bCIs a parameter to be trained;
the associated gate weights are as follows:
Γr=σ(Wr[ht-1,xt]T+br)
in the formula, WrAnd brIs the parameter to be trained.
The word-level encoder word encoder specifically comprises:
xit=Wewit,t∈[1,T]
Figure RE-GDA0002318502020000064
Figure RE-GDA0002318502020000065
in the formula, xitFor the result of vectorization of the integer number corresponding to a word, WeIs an embedded matrix, witIs the vector after the integer number one-hot coding OneHot in words, T is the maximum value of the word number of each sentence,
Figure RE-GDA0002318502020000066
andthe outputs of the GRUs in forward and reverse directions, respectively.
The syntax level encoder transmit encoder is specifically:
Figure RE-GDA0002318502020000071
in the formula, siIs the presence _ entry of the ith sentence, and L is the number of sentences.
The word-level attention layer word attention specifically comprises:
uit=tanh(Wwhit+bw)
Figure RE-GDA0002318502020000073
Figure RE-GDA0002318502020000074
in the formula, hitBy passing
Figure RE-GDA0002318502020000075
And
Figure RE-GDA0002318502020000076
is spliced to obtain siIs the sense _ entry, W, of the ith sentencewAnd bwIs the parameter to be trained.
The sentence-level attention layer sensor _ attribute specifically includes:
ui=tanh(Wshi+bs)
Figure RE-GDA0002318502020000077
Figure RE-GDA0002318502020000078
in the formula, hiBy passingAndis spliced to obtainsAnd bsIs the parameter to be trained.
(3) Language model framework: as shown in fig. 2, word vectors corresponding to doc _ attention and words of the hierarchical attention network are spliced and mapped to label, and the mapping method uses softmax; the word vector is an embedded matrix WeBy means of the integer digits of words, can be embedded from the matrix WeIndex to its corresponding word vector.
(4) The hierarchical attention network and the language model framework form a language model, the language model is trained, parameters to be trained in the model are fitted, and doc _ attention corresponding to each document is a final document vector after the model is trained; for the prediction data, because the parameters to be trained in the model are fitted, the document vector doc _ attention can also be obtained; the obtained document vector can be used in natural language processing tasks of text classification, text similarity calculation and emotion analysis.
The invention also carries out document vector training based on the language model, but obtains the document vector by using the hierarchical attention network instead of direct training, so that the model can pay attention to the information of the whole document, and the use cost is greatly reduced.
The above-described embodiments are intended to illustrate rather than to limit the invention, and any modifications and variations of the present invention are within the spirit of the invention and the scope of the appended claims.

Claims (10)

1. A method for generating a document vector, the method comprising the steps of:
(1) preparing data;
(1.1) converting the words of each sentence in the document into a list, forming the list of the document by the list of each sentence, and finally mapping each word in the document list to a unique integer number which is recorded as doc.
(1.2) sliding a window with the window size of window _ size in a document list, recording a word list in the window as words, and recording the second window _ size//2 words in the window as label; one document will eventually form a plurality of words and labels, while the data of one sample will include doc, words and label, and one document will eventually form a plurality of samples as training samples of the hierarchical attention network, which samples have the same doc, but different words and labels.
(2) Constructing a hierarchical attention network; the hierarchical attention network comprises a word-level encoder word encoder, a word-level attention layer word attention, a sentence-level encoder sensor encoder, a sentence-level attention layer sensor attention and a document attention doc attention; the word encoder layer and the sensor encoder layer are both of a sequence encoder layer type; the word attribute layer and the presence attribute layer are attention networks, the context text context of the word attribute layer and the presence attribute layer are different, and the context of the word attribute layer is uwSen, thenContext at tenceattention level is us
(3) Language model framework: splicing word vectors corresponding to doc _ attribute and words output by the hierarchical attention network, and mapping to label by adopting softmax; the word vector is an embedded matrix WeBy the integer number of words from the embedded matrix WeIndex to its corresponding word vector.
(4) The hierarchical attention network and the language model framework form a language model, the language model is trained, parameters to be trained in the model are fitted, and doc _ attention corresponding to each document is a final document vector after the language model is trained; the prediction data is input into the model, and a document vector of the prediction data can be obtained.
2. A method for generating document vectors according to claim 1, wherein in step (1.1), words of each sentence in the document are converted into a list, the list of each sentence constitutes the list of the document, and finally each word in the document list is mapped to a unique integer number.
3. The method according to claim 1, wherein in step (1.2), both words and label are the result of mapping words to integer numbers.
4. The method of claim 1, wherein in step (1.2), the window size is 8.
5. The method of claim 1, wherein in step (1.2), window _ size//2 represents the integer portion of the quotient of window size divided by 2.
6. The method of claim 1, wherein in step (2), the sequence encoder is based on a GRU (gated Current Unit) network, and the GRU updates the hidden key mainly through the associative gate and the refresh gateThe hidden state h is the hidden state of the current step ttThen the update gate is:
Figure FDA0002248462780000021
wherein the content of the first and second substances,
Γu=σ(Wu[ht-1,xt]T+bu)
where σ is the sigmoid function, WuAnd buIs a parameter to be trained and is,to associate the door, xtIs the input of the GRU;
the associated gates are as follows:
wherein tanh represents a hyperbolic tangent function, WCAnd bCIs a parameter to be trained;
the associated gate weights are as follows:
Γr=σ(Wr[ht-1,xt]T+br)
in the formula, WrAnd brIs the parameter to be trained.
7. The method for generating document vectors according to claim 1, wherein in the step (2), the word-level encoder word encoder specifically comprises:
xit=Wewit,t∈[1,T]
Figure FDA0002248462780000024
Figure FDA0002248462780000025
in the formula, xitFor the result of vectorization of the integer number corresponding to a word, WeIs an embedded matrix, witIs the vector after the integer number one-hot coding OneHot in words, T is the maximum value of the word number of each sentence,
Figure FDA0002248462780000026
andthe outputs of the GRUs in forward and reverse directions, respectively.
8. The method for generating document vectors according to claim 1, wherein in the step (2), the syntax-level encoder sensor encoder is specifically:
Figure FDA0002248462780000028
Figure FDA0002248462780000029
in the formula, siIs the presence _ entry of the ith sentence, and L is the number of sentences.
9. The method for generating document vectors according to claim 1, wherein in the step (2), the word-level attention layer word attention specifically comprises:
uit=tanh(Wwhit+bw)
Figure FDA0002248462780000032
in the formula, hitBy passing
Figure FDA0002248462780000033
Andis spliced to obtain siIs the sense _ entry, W, of the ith sentencewAnd bwIs the parameter to be trained.
10. The method for generating document vectors according to claim 1, wherein in step (2), the sentence-level attention level sensor _ attention specifically comprises:
ui=tanh(Wshi+bs)
Figure FDA0002248462780000036
in the formula, hiBy passing
Figure FDA0002248462780000037
And
Figure FDA0002248462780000038
is spliced to obtainsAnd bsIs the parameter to be trained.
CN201911025383.7A 2019-10-25 2019-10-25 Document vector generation method Pending CN110852070A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201911025383.7A CN110852070A (en) 2019-10-25 2019-10-25 Document vector generation method

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201911025383.7A CN110852070A (en) 2019-10-25 2019-10-25 Document vector generation method

Publications (1)

Publication Number Publication Date
CN110852070A true CN110852070A (en) 2020-02-28

Family

ID=69597863

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201911025383.7A Pending CN110852070A (en) 2019-10-25 2019-10-25 Document vector generation method

Country Status (1)

Country Link
CN (1) CN110852070A (en)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN112230990A (en) * 2020-11-10 2021-01-15 北京邮电大学 Program code duplication checking method based on hierarchical attention neural network

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN108153864A (en) * 2017-12-25 2018-06-12 北京牡丹电子集团有限责任公司数字电视技术中心 Method based on neural network generation text snippet
CN108460019A (en) * 2018-02-28 2018-08-28 福州大学 A kind of emerging much-talked-about topic detecting system based on attention mechanism
CN108829818A (en) * 2018-06-12 2018-11-16 中国科学院计算技术研究所 A kind of file classification method
CN109145304A (en) * 2018-09-07 2019-01-04 中山大学 A kind of Chinese Opinion element sentiment analysis method based on word

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN108153864A (en) * 2017-12-25 2018-06-12 北京牡丹电子集团有限责任公司数字电视技术中心 Method based on neural network generation text snippet
CN108460019A (en) * 2018-02-28 2018-08-28 福州大学 A kind of emerging much-talked-about topic detecting system based on attention mechanism
CN108829818A (en) * 2018-06-12 2018-11-16 中国科学院计算技术研究所 A kind of file classification method
CN109145304A (en) * 2018-09-07 2019-01-04 中山大学 A kind of Chinese Opinion element sentiment analysis method based on word

Non-Patent Citations (3)

* Cited by examiner, † Cited by third party
Title
QUOC LE等: "Distributed Representations of Sentences and Documents", 《PROCEEDINGS OF THE 31ST INTERNATIONAL CONFERENCE ON MACHINE LEARNING》 *
ZICHAO YANG等: "Hierarchical Attention Networks for Document Classification", 《PROCEEDINGS OF NAACL-HLT 2016》 *
欧阳文俊等: "基于层级注意力模型的无监督文档表示学习", 《计算机系统应用》 *

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN112230990A (en) * 2020-11-10 2021-01-15 北京邮电大学 Program code duplication checking method based on hierarchical attention neural network

Similar Documents

Publication Publication Date Title
CN110287320B (en) Deep learning multi-classification emotion analysis model combining attention mechanism
CN108984724B (en) Method for improving emotion classification accuracy of specific attributes by using high-dimensional representation
CN110929030B (en) Text abstract and emotion classification combined training method
CN112214599B (en) Multi-label text classification method based on statistics and pre-training language model
CN109740148B (en) Text emotion analysis method combining BiLSTM with Attention mechanism
CN111209738B (en) Multi-task named entity recognition method combining text classification
CN110287323B (en) Target-oriented emotion classification method
CN112800776B (en) Bidirectional GRU relation extraction data processing method, system, terminal and medium
Zhang et al. Multi-layer attention based CNN for target-dependent sentiment classification
CN110472042B (en) Fine-grained emotion classification method
CN110765260A (en) Information recommendation method based on convolutional neural network and joint attention mechanism
CN108062388A (en) Interactive reply generation method and device
CN111079409B (en) Emotion classification method utilizing context and aspect memory information
CN107818084B (en) Emotion analysis method fused with comment matching diagram
CN110659411B (en) Personalized recommendation method based on neural attention self-encoder
CN110580287A (en) Emotion classification method based ON transfer learning and ON-LSTM
CN113704546A (en) Video natural language text retrieval method based on space time sequence characteristics
CN111597340A (en) Text classification method and device and readable storage medium
CN113705238B (en) Method and system for analyzing aspect level emotion based on BERT and aspect feature positioning model
CN111538841B (en) Comment emotion analysis method, device and system based on knowledge mutual distillation
CN111368082A (en) Emotion analysis method for domain adaptive word embedding based on hierarchical network
CN114358201A (en) Text-based emotion classification method and device, computer equipment and storage medium
Kalaivani et al. A review on feature extraction techniques for sentiment classification
CN115408603A (en) Online question-answer community expert recommendation method based on multi-head self-attention mechanism
CN116680363A (en) Emotion analysis method based on multi-mode comment data

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
RJ01 Rejection of invention patent application after publication
RJ01 Rejection of invention patent application after publication

Application publication date: 20200228