CN114757183B - Cross-domain emotion classification method based on comparison alignment network - Google Patents

Cross-domain emotion classification method based on comparison alignment network Download PDF

Info

Publication number
CN114757183B
CN114757183B CN202210373901.XA CN202210373901A CN114757183B CN 114757183 B CN114757183 B CN 114757183B CN 202210373901 A CN202210373901 A CN 202210373901A CN 114757183 B CN114757183 B CN 114757183B
Authority
CN
China
Prior art keywords
domain
cross
target
original
field
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN202210373901.XA
Other languages
Chinese (zh)
Other versions
CN114757183A (en
Inventor
宋大为
马放
张辰
杨艺
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Beijing Institute of Technology BIT
Original Assignee
Beijing Institute of Technology BIT
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Beijing Institute of Technology BIT filed Critical Beijing Institute of Technology BIT
Priority to CN202210373901.XA priority Critical patent/CN114757183B/en
Publication of CN114757183A publication Critical patent/CN114757183A/en
Application granted granted Critical
Publication of CN114757183B publication Critical patent/CN114757183B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F40/00Handling natural language data
    • G06F40/20Natural language analysis
    • G06F40/279Recognition of textual entities
    • G06F40/289Phrasal analysis, e.g. finite state techniques or chunking
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/30Information retrieval; Database structures therefor; File system structures therefor of unstructured textual data
    • G06F16/35Clustering; Classification
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F40/00Handling natural language data
    • G06F40/30Semantic analysis

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Engineering & Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Audiology, Speech & Language Pathology (AREA)
  • Computational Linguistics (AREA)
  • General Health & Medical Sciences (AREA)
  • Health & Medical Sciences (AREA)
  • Artificial Intelligence (AREA)
  • Data Mining & Analysis (AREA)
  • Databases & Information Systems (AREA)
  • Information Retrieval, Db Structures And Fs Structures Therefor (AREA)
  • Machine Translation (AREA)

Abstract

The invention relates to a cross-domain emotion analysis method based on a comparison alignment network, and belongs to the technical field of fine-granularity emotion analysis in natural language processing. The invention researches an underexplored scene of cross-domain emotion classification, namely a scene with a few samples in the target domain. In this scenario, the present invention proposes a neural network model named Contrast Alignment Network (CAN). The model first randomly extracts two instances from the original domain and the target domain, and then trains them according to the instances of the combined target domain and the original domain. The first objective is to minimize classification errors on the original domain. The second is a pairwise comparison objective, where the distance measure between the objective domain instance and the original domain instance in a pair is minimized if they express the same emotion, otherwise the measure is maximized at a constant upper limit. The method solves the problem of limited data resources in the target field in the cross-field emotion classification task, and improves the use experience of users.

Description

Cross-domain emotion classification method based on comparison alignment network
Technical Field
The invention relates to a cross-domain emotion classification method, in particular to a cross-domain emotion analysis method based on a comparison alignment network, and belongs to the technical field of fine-granularity emotion analysis in natural language processing.
Background
Cross-domain emotion classification (Cross Domain Sentiment Classification, CDSC) is an important task aimed at transferring learned knowledge from the original domain to the target domain. CDSC enables emotion classification models trained in the original domain with a large amount of labeled data to perform well in target domain data with limited training samples. This situation is common in the industry when the data of the target domain is lacking and the data of the original domain is sufficient, and has a challenge, the main challenge being domain transfer (or distribution transfer) between the source domain and the target domain. The domain transfer problem is mainly a distribution difference between any two domains, for example, a word used in the medical domain is greatly different from a word used in the restaurant domain.
Domain transfer is an important problem in cross-domain emotion classification, which can be alleviated to a large extent by domain adaptation methods. Currently, researchers have proposed various field-adaptive models. These models require a large amount of unlabeled data from the target domain so that they can learn a good representation of each target instance as the correct input to the classifier trained in the original domain.
Currently, unsupervised Domain Adaptation (UDA) techniques have been used to solve the domain transfer problem. Essentially, the UDA utilizes other unlabeled data in the target domain to minimize domain offset by aligning statistics across domains. However, in practical applications, a large amount of unlabeled target area data required for UDA may not be sufficiently available, thereby limiting the applicability of UDA.
Meanwhile, there is a scene which is not fully explored at present in cross-domain emotion classification, namely, a scene with a few samples in the target domain, and the scene exists in many practical applications in the industry. Unlike an unsupervised scenario, a few sample scenario does not require additional unlabeled target domain data, but relies only on the scarce labeling data available in the target domain.
Disclosure of Invention
The invention aims to overcome the defects of the prior art, creatively provides a cross-domain emotion classification method based on a comparison alignment network, and aims to solve the technical problem of data resource limitation in the target domain. The invention researches an underexplored scene of cross-domain emotion classification, namely a scene with a few samples in the target domain. Unlike an unsupervised scenario, a few sample scenario does not require additional unlabeled target domain data, but relies only on the scarce labeling data available in the target domain. In this scenario, the present invention proposes a neural network model named Contrast Alignment Network (CAN). The model first randomly extracts two instances from the original domain and the target domain, and then trains them according to the instances of the combined target domain and the original domain. The first objective is to minimize classification errors on the original domain. The second is a pairwise comparison objective, where the distance measure between the objective domain instance and the original domain instance in a pair is minimized if they express the same emotion, otherwise the measure is maximized at a constant upper limit.
Notably, the motivation for the alignment of networks is not just to address resource limitation issues. In fact, by deducing that the model provided by the invention has more complete learning limit, theoretical basis is provided for the basic efficacy of the model.
Cross-domain emotion analysis CDSC requires an in-situ domainModel of middle training, original fieldContains n s examples, which are specific to the target field/>And performing emotion classification. X i s is the ith sample sentence in the original field, and y i s is the label corresponding to the ith sample in the original field. In other words, the original field/>In example sample distribution compliance/>Also, there is/>, on the target areaX t represents a sample sentence of the target field t, and y t represents a tag corresponding to the sample sentence of the target field t. /(I)And/>Representing the domain offset between the joint probabilities of the source and target, respectively, and
The concept of domain adaptation is imposed to alleviate the domain transfer problem. In the case that there is insufficient tagged data in the target domain, the unsupervised domain adaptation method uses a large amount of additional untagged target dataTo align the two fields, n t being the target field and the number of marked data samples. However, the large amount of unlabeled data required may not always be available in the target area. Thus, the method focuses on a more realistic case of cross-domain emotion analysis, i.e., a small sample case, instead using a small number of auxiliary markers to target domain data/>N t<<ns,nt is the number of auxiliary mark target field data samples.
Inspired by the adaptation work of the existing unsupervised field, the learning target is decomposed into two parts: and calculating the discriminative original domain risk and regularized domain transfer. In order to further minimize the risk of the original domain, the invention provides a method for comparing and aligning the original domain and the target domain by taking example classification information as a condition.
The technical mode adopted by the invention is as follows.
An emotion classification method based on a comparison alignment network comprises the following steps:
First, text data preprocessing is performed.
And loading a comment corpus and a pre-training language model, and carrying out text preprocessing and text data formatting on comment text data in the comment corpus. Wherein, the pre-training language model can adopt BERT model, roBERTa model and the like.
Then, constructing a cross-domain emotion classification model based on the comparison alignment network.
The cross-domain emotion classification model f based on the contrast alignment network provided by the invention comprises an encoder g θ and a classifier h φ, and on the basis of the architecture, an original domain classification target loss function is introducedAnd comparing the target domain classification target loss function/>
Wherein the encoder g θ uses the pre-trained language model as a basis for encoding context information of comment sentences. Preferably, the encoder uses CLS (sentence vector representation) full sentence representation of the pre-trained language model as the context hidden state representation vector H of the whole comment sentence, h= { H 1,h2,...,hn},hn representing the hidden state representation vector of the nth token.
The classifier h φ consists of a multi-layer perceptron MLP and a softmax layer (soft maximization normalization layer). Wherein, multilayer perceptron includes four layers, does in proper order: a full connection layer, a ReLU (Linear rectifier function) activation function layer, a dropout layer (random discard layer) and a full connection layer. The output representation through the MLP is sent to the softmax layer (soft maximization normalization layer), from which the corresponding losses are calculated.
Then, the discriminative primitive domain risk is calculated.
In the method, for the discriminative original domain risk, an empirical classification loss term of the original domain is adopted, and a classification target is modeled as a loss based on cross entropy
Wherein n s is the number of data samples in the original field; y i is the label of the ith sample of the original field data,And (5) predicting labels on the ith sample of the original field data for the model.
Then, the original domain and the target domain are aligned by comparing the example-level classification information.
The goal is to minimize domain offset under limited target domain data. Although alignment of distribution levels is difficult to achieve with limited target data, alignment of instance levels may be combined and alignment of instance levels may achieve distribution level alignment with some probability.
In the method of the present invention, contrast loss is introduced to minimize the distance metric between the sample pair (if they share the same emotion) between the original domain instance and the target domain instance, otherwise the metric is maximized. Therefore, regardless of the field, emotion tags can be accurately assigned. Although the distance measure between the original domain and the target domain can be directly applied to the input x, the present invention follows most domain-adaptive methods to align the hidden representation Z, which will be more abstract and can capture more semantic information.
Specifically, given any pair ofSpecific contrast loss/>The calculation is as follows:
Wherein, X i s represents the ith sample comment sentence in the original field, X j t represents the jth sample comment sentence in the target field, y i s represents the label corresponding to the ith sample comment sentence in the original field, and y j t represents the label corresponding to the jth sample comment sentence in the target field; representing a distance measure between the original domain instance and the target domain instance; /(I) Indicating a function for the equation; m is a predefined constant.
The contrast loss objective pushes the instances of the same emotion polarity closer, rejecting the instances of different emotion polarity. Furthermore, the goal is not to push different clusters to infinity, but rather to limit the scope of rejection to a constant as a slack in the learning algorithm.
And then, carrying out regularized field transfer.
The overall objective of the method includes cross entropy loss function of the original domain dataComparing and aligning original domain and target domain loss function/>And pass/>Regularization minimizes cross entropy loss function/>, of raw domain dataAnd comparing the aligned original domain with the target domain loss function/>
Wherein the overall objective function is:
Where α is a trade-off term between classification and comparison targets and λ is the regularization coefficient of all model parameters Θ= { θ, Φ }.
Then, model training is performed. For the overall objective function using a standard batch random gradient descent algorithmTraining is performed.
Specifically, batch iterative training is carried out on all training samples in a training set, and a trained cross-domain emotion classification model based on a comparison alignment network is obtained.
And finally, performing cross-domain emotion classification by using a trained cross-domain emotion classification model based on the comparison and alignment network.
Advantageous effects
Compared with the prior art, the method has the following advantages:
1. The method solves the problem that data resources in the target field are limited in the cross-field emotion classification task. The invention researches an underexplored scene of cross-domain emotion classification, namely a scene with a few samples in the target domain. Unlike an unsupervised scenario, a few sample scenario does not require additional unlabeled target domain data, but relies only on the scarce labeling data available in the target domain.
2. The neural network model for the alignment network proposed by the method comprises two targets, wherein the first target is to minimize classification errors in the original field. The second is a pairwise comparison objective, where the distance measure between the objective domain instance and the original domain instance in a pair is minimized if they express the same emotion, otherwise the measure is maximized at a constant upper limit.
3. The contrast alignment network model provided by the invention has the advantage that the performance of the contrast alignment network model on the cross-domain emotion classification task is obviously superior to that of a corresponding baseline model. The performance of other baseline models on cross-domain datasets drops dramatically. In contrast, the proposed contrast aligned neural network model is more robust than other baseline models.
4. The contrast alignment neural network model provided by the invention not only solves the problem of limited resources in the target field, but also has a more complete learning limit, thereby providing a theoretical basis for the basic efficacy of the model.
5. The method improves the problems in cross-domain emotion classification of the existing fine-grained emotion analysis, and can well improve the use experience of users.
Drawings
FIG. 1 is an overall flow chart of the method of the present invention.
FIG. 2 is an example diagram of a cross-domain emotion classification task.
Fig. 3 is a visual representation of the effect of the method of the present invention.
Detailed Description
The invention is described in further detail below with reference to the drawings and examples.
Examples
As shown in fig. 1, a cross-domain emotion classification method based on a comparison alignment network includes the following steps:
Step 1: text preprocessing.
First, a comment corpus and a pre-trained language model are loaded. The pre-training language model may be a BERT model, or may be another model (such as RoBERTa models).
Then, text preprocessing and text data formatting are performed on the comment corpus.
Specifically, the method comprises the following steps:
Step 1.1: and extracting attribute words, viewpoint words and position information of the attribute words and the viewpoint words from each comment sentence.
Step 1.2: and pre-segmenting the comment sentences by using a nltk word segmentation device, and separating the segmented token words by using spaces.
Step 1.3: after the comment sentence is divided into a token sequence, two special token words are added: [ CLS ], [ SEP ], thereby constructed in the general input form: s= { [ CLS ], w 1,w2...,wn, [ SEP ] }, n represents the total number of token words of the comment sentence, and w n represents the n-th token of the comment sentence.
Step 1.4: text data formatting is performed.
And (3) carrying out filling processing on each comment sentence token word sequence to ensure that the length of each comment sentence token word sequence is 128. Each token word in the comment sentence is subjected to tokenize operations using the word segmenter tokenizer of the pre-trained language model. The processed data set is divided into a training set, a validation set and a test set, and a batch data form is constructed.
Step 2: and constructing a cross-domain emotion classification model based on the comparison alignment network.
The cross-domain emotion classification model f based on the comparison alignment network consists of an encoder g θ and a classifier h φ. Based on encoder g θ and classifier h φ architecture, the original domain classification target loss function is introducedAnd comparing the target domain classification target loss function/>
Wherein the encoder g θ encodes context information of the comment sentence. For emotion classification tasks, it is important that emotion information of a sentence is contained in the context of the sentence and that the entire sentence is context modeled. Thus, using the full sentence CLS representation of the pre-trained language model, the context hidden state representation vector H, h= { H 1,h2,...,hn},hn, represents the hidden state representation vector of the nth token as the entire comment sentence. Each token word in the sequence of comment sentences is mapped to a code vector.
The classifier h φ consists of a multi-layer perceptron MLP and a softmax layer (soft maximization normalization layer). Wherein, the multilayer perceptron includes four layers: full connection layer, reLU (linear rectification function) activation function layer, dropout layer (random discard layer) and full connection layer. The output representation through the fully-connected layer is sent to the softmax layer (soft maximization normalization layer), whereby the corresponding label is predicted and the corresponding target loss is calculated.
Step 3: and calculating the risk of the discriminant primitive field.
For the discriminative primitive domain risk, the invention adopts the empirical classification loss term of the primitive domain, and then models the classification target as the loss based on cross entropy
Wherein n s is the number of samples of the original field data, y i is the label of the ith sample of the original field data,And (5) predicting labels on the ith sample of the original field data for the model.
Step 4: and comparing and aligning the original domain with the target domain by using example-level classification information.
The domain offset is minimized under limited target domain data. Although alignment of distribution levels is difficult to achieve with limited target data, alignment of instance levels may be combined and alignment of instance levels may achieve distribution level alignment with some probability.
Thus, the present invention introduces contrast loss to minimize the distance metric between the sample pair original domain instance and the target domain instance if the original domain instance and the target domain instance share the same emotion, and otherwise maximize the metric. Regardless of the field, emotion tags can be accurately assigned.
Although the distance measure between the original domain and the target domain can be directly applied to the input x, the present invention follows most domain-adaptive methods, aligning the hidden representation Z. This is more abstract and more semantic information can be captured.
Specifically, given any pair ofThe specific contrast loss is calculated as:
Wherein, Representing a distance measure between the original domain instance and the target domain instance. /(I)The function is indicated for the equation. m is a predefined constant.
The contrast loss objective pushes the instances of the same emotion polarity closer, rejecting the instances of different emotion polarity. Furthermore, the goal is not to push different clusters to infinity, but rather to limit the scope of rejection to a constant as a slack in the learning algorithm.
Step 5: regularized domain transfer.
The overall objective of the method includes cross entropy loss function of the original domain dataComparing and aligning original domain and target domain loss function/>By/>Regularization minimizes cross entropy loss function/>, of raw domain dataAnd comparing the aligned original domain with the target domain loss function/>
The overall objective function is:
Where α is a trade-off term between classification and comparison targets, λ is a regularization coefficient of all model parameters Θ= { θ, Φ } θ represents a parameter of the encoder, Φ represents a parameter of the classifier.
Step 6: training a cross-domain emotion classification model based on a comparison alignment network.
For the overall objective function using a standard batch random gradient descent algorithmAnd (5) performing optimization training.
Specifically, batch iterative training is carried out on all training samples in a training set, and a trained cross-domain emotion classification model based on a comparison alignment network is obtained.
Step 7: and performing cross-domain emotion classification by using the trained cross-domain emotion classification model based on the comparison and alignment network.
Further, the method can be evaluated. After the training is completed in the training set, a verification test is performed in the verification set used. The evaluation indexes used include:
for cross-domain emotion classification, using accuracy and F1 values as evaluation indexes;
And updating the optimal model for each round of verification and saving.
Test verification
The method was tested. Firstly, loading the parameters and test data of the optimal model stored before, and then converting the test data into a required format and inputting the required format into the optimal model for testing. Wherein, the evaluation index is the same as the evaluation index used in verification.
As shown in fig. 2, for the original field being the comment field of the notebook computer, the comment sentence "the brand notebook computer is too stuck, the use experience is extremely bad", the target field being the comment sentence of the restaurant field "the cooking of his home is general, the service is also general", and there is a fine-grained field deviation between the original field and the target field data, which makes most of the inter-field emotion classification methods and the unsupervised self-adaptive methods unable to perform another migration well, and cannot effectively judge the emotion polarity of the comment sentence of the target field on the target field.
As shown in fig. 3, the cross-domain (notebook computer comment field and restaurant comment field) effect visualization representation of the emotion classification method based on the contrast alignment network, the left graph is the cross-domain effect visualization of the non-contrast alignment network, and the right graph is the cross-domain effect visualization of the contrast alignment network. It can be seen that the model effect visualization with right graph with contrast aligned network has a more elegant manifold than the effect visualization without contrast aligned network, which means that the contrast aligned network has generalization capability even if the target data is small.
The foregoing is a preferred embodiment of the present invention and the present invention should not be limited to the embodiment and the disclosure of the drawings. All equivalents and modifications that come within the spirit of the disclosure are desired to be protected.

Claims (5)

1. The cross-domain emotion classification method based on the comparison alignment network is characterized by comprising the following steps of:
step 1: loading a comment corpus and a pre-training language model, and carrying out text preprocessing and text data formatting on comment text data in the comment corpus;
step 2: constructing a cross-domain emotion classification model based on a comparison alignment network;
The cross-domain emotion classification model f based on the comparison alignment network consists of an encoder g θ and a classifier h φ, and an original domain classification target loss function is introduced on the basis of the framework And comparing the target domain classification target loss function/>
The encoder g θ uses the pre-trained language model as a basis for encoding context information of comment sentences; the classifier h φ consists of a multi-layer perceptron MLP and a softmax layer; passing the output representation through the multi-layer perceptron to a softmax layer, thereby calculating a corresponding loss;
step 3: calculating the risk of the discriminant primitive field;
for discriminative primitive domain risk, modeling a classification target as a cross entropy-based loss by using an empirical classification loss term of the primitive domain
Wherein n s is the number of data samples in the original field; y i is the label of the ith sample of the original field data,A predictive label of the model on an ith sample of the original field data;
step 4: comparing and aligning the original field with the target field according to the example-level classification information;
Given any pair of Specific contrast loss/>The calculation is as follows:
wherein, X i s represents the ith sample comment sentence in the original field, X j t represents the jth sample comment sentence in the target field, y i s represents the label corresponding to the ith sample comment sentence in the original field, and y i t represents the label corresponding to the jth sample comment sentence in the target field; representing a distance measure between the original domain instance and the target domain instance; /(I) Indicating a function for the equation; m is a predefined constant;
step 5: performing regularization field transfer; the overall objective includes a cross entropy loss function of the raw domain data Comparing and aligning original domain and target domain loss function/>And pass/>Regularization minimizes cross entropy loss function of raw domain dataAnd comparing the aligned original domain with the target domain loss function/>
Wherein the overall objective function is:
where α is a trade-off term between classification and comparison targets, λ is the regularization coefficient of all model parameters Θ= { θ, Φ };
step 6: for the overall objective function using a standard batch random gradient descent algorithm Training to obtain a trained cross-domain emotion classification model based on a comparison alignment network;
step 7: and performing cross-domain emotion classification by using the trained cross-domain emotion classification model based on the comparison and alignment network.
2. The cross-domain emotion classification method based on a contrast alignment network of claim 1, wherein the encoder uses CLS full sentence representation of the pre-trained language model as context hidden state representation vector H of the whole comment sentence, h= { H 1,h2,…,hn},hn represents hidden state representation vector of the nth token.
3. The cross-domain emotion classification method based on a comparison alignment network of claim 1, wherein in step 2, the multi-layer perceptron comprises four layers, in order: a full connection layer, a ReLU activation function layer, a dropout layer and a full connection layer.
4. The cross-domain emotion classification method based on a comparison alignment network of claim 1, wherein step 1 comprises the steps of:
step 1.1: extracting attribute words, viewpoint words and position information of the attribute words and the viewpoint words from each comment sentence;
step 1.2: pre-segmenting the comment sentences by using nltk word segmentation devices, and separating segmented token words by using spaces;
Step 1.3: after the comment sentence is divided into a token sequence, two special token words are added: [ CLS ], [ SEP ], thereby constructed in the general input form: s= { [ CLS ], w 1,w2…,wn, [ SEP ] }, n represents the total number of token words of the comment sentence, and w n represents the n-th token of the comment sentence;
Step 1.4: formatting text data;
Performing filling processing on each comment sentence token word sequence to ensure that the length of each comment sentence token word sequence is 128; using a word segmentation device tokenizer of the pre-training language model to carry out tokenize operation on each token word in the comment sentence; the processed data set is divided into a training set, a validation set and a test set, and a batch data form is constructed.
5. The cross-domain emotion classification method based on a comparison alignment network of claim 1, wherein after training is completed in the training set, a verification test is performed in the verification set used, and the evaluation index includes:
for cross-domain emotion classification, using accuracy and F1 values as evaluation indexes;
Updating and saving the optimal model for each round of verification.
CN202210373901.XA 2022-04-11 2022-04-11 Cross-domain emotion classification method based on comparison alignment network Active CN114757183B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202210373901.XA CN114757183B (en) 2022-04-11 2022-04-11 Cross-domain emotion classification method based on comparison alignment network

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202210373901.XA CN114757183B (en) 2022-04-11 2022-04-11 Cross-domain emotion classification method based on comparison alignment network

Publications (2)

Publication Number Publication Date
CN114757183A CN114757183A (en) 2022-07-15
CN114757183B true CN114757183B (en) 2024-05-10

Family

ID=82328537

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202210373901.XA Active CN114757183B (en) 2022-04-11 2022-04-11 Cross-domain emotion classification method based on comparison alignment network

Country Status (1)

Country Link
CN (1) CN114757183B (en)

Families Citing this family (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN115239937B (en) * 2022-09-23 2022-12-20 西南交通大学 Cross-modal emotion prediction method

Citations (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109308318A (en) * 2018-08-14 2019-02-05 深圳大学 Training method, device, equipment and the medium of cross-domain texts sentiment classification model
CN109753566A (en) * 2019-01-09 2019-05-14 大连民族大学 The model training method of cross-cutting sentiment analysis based on convolutional neural networks
CN110032646A (en) * 2019-05-08 2019-07-19 山西财经大学 The cross-domain texts sensibility classification method of combination learning is adapted to based on multi-source field
CN110503092A (en) * 2019-07-22 2019-11-26 天津科技大学 The improvement SSD monitor video object detection method adapted to based on field
AU2020100710A4 (en) * 2020-05-05 2020-06-11 Chen, Dadu Mr A method for sentiment analysis of film reviews based on deep learning and natural language processing
CN112163091A (en) * 2020-09-25 2021-01-01 大连民族大学 CNN-based aspect-level cross-domain emotion analysis method
CN113326378A (en) * 2021-06-16 2021-08-31 山西财经大学 Cross-domain text emotion classification method based on parameter migration and attention sharing mechanism

Patent Citations (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109308318A (en) * 2018-08-14 2019-02-05 深圳大学 Training method, device, equipment and the medium of cross-domain texts sentiment classification model
CN109753566A (en) * 2019-01-09 2019-05-14 大连民族大学 The model training method of cross-cutting sentiment analysis based on convolutional neural networks
CN110032646A (en) * 2019-05-08 2019-07-19 山西财经大学 The cross-domain texts sensibility classification method of combination learning is adapted to based on multi-source field
CN110503092A (en) * 2019-07-22 2019-11-26 天津科技大学 The improvement SSD monitor video object detection method adapted to based on field
AU2020100710A4 (en) * 2020-05-05 2020-06-11 Chen, Dadu Mr A method for sentiment analysis of film reviews based on deep learning and natural language processing
CN112163091A (en) * 2020-09-25 2021-01-01 大连民族大学 CNN-based aspect-level cross-domain emotion analysis method
CN113326378A (en) * 2021-06-16 2021-08-31 山西财经大学 Cross-domain text emotion classification method based on parameter migration and attention sharing mechanism

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
基于集成深度迁移学习的多源跨领域情感分类;赵传君;王素格;李德玉;;山西大学学报(自然科学版);20180404(第04期);全文 *

Also Published As

Publication number Publication date
CN114757183A (en) 2022-07-15

Similar Documents

Publication Publication Date Title
CN112214995B (en) Hierarchical multitasking term embedded learning for synonym prediction
WO2023065545A1 (en) Risk prediction method and apparatus, and device and storage medium
CN111738003B (en) Named entity recognition model training method, named entity recognition method and medium
CN111737552A (en) Method, device and equipment for extracting training information model and acquiring knowledge graph
CN112541355B (en) Entity boundary type decoupling few-sample named entity recognition method and system
CN112084327A (en) Classification of sparsely labeled text documents while preserving semantics
CN111985239A (en) Entity identification method and device, electronic equipment and storage medium
CN111506732B (en) Text multi-level label classification method
CN111597340A (en) Text classification method and device and readable storage medium
CN113254675B (en) Knowledge graph construction method based on self-adaptive few-sample relation extraction
CN112100348A (en) Knowledge base question-answer relation detection method and system of multi-granularity attention mechanism
CN114912423A (en) Method and device for analyzing aspect level emotion based on transfer learning
CN114841151B (en) Medical text entity relation joint extraction method based on decomposition-recombination strategy
CN116383399A (en) Event public opinion risk prediction method and system
CN115063119A (en) Recruitment decision system and method based on adaptivity of recruitment behavior data
CN113704392A (en) Method, device and equipment for extracting entity relationship in text and storage medium
CN115935991A (en) Multitask model generation method and device, computer equipment and storage medium
CN113591955A (en) Method, system, equipment and medium for extracting global information of graph data
CN115563314A (en) Knowledge graph representation learning method for multi-source information fusion enhancement
CN116861924A (en) Project risk early warning method and system based on artificial intelligence
CN114757183B (en) Cross-domain emotion classification method based on comparison alignment network
CN115659947A (en) Multi-item selection answering method and system based on machine reading understanding and text summarization
CN115017879A (en) Text comparison method, computer device and computer storage medium
CN114780723A (en) Portrait generation method, system and medium based on guide network text classification
CN114582448A (en) Epidemic case information extraction framework construction method based on pre-training language model

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant