CN111813894A - Natural language emotion recognition method based on deep learning - Google Patents
Natural language emotion recognition method based on deep learning Download PDFInfo
- Publication number
- CN111813894A CN111813894A CN202010613189.7A CN202010613189A CN111813894A CN 111813894 A CN111813894 A CN 111813894A CN 202010613189 A CN202010613189 A CN 202010613189A CN 111813894 A CN111813894 A CN 111813894A
- Authority
- CN
- China
- Prior art keywords
- layer
- text
- model
- picture
- data
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Pending
Links
- 230000008909 emotion recognition Effects 0.000 title claims abstract description 39
- 238000000034 method Methods 0.000 title claims abstract description 24
- 238000013135 deep learning Methods 0.000 title claims abstract description 13
- 238000012549 training Methods 0.000 claims abstract description 57
- 238000013527 convolutional neural network Methods 0.000 claims abstract description 42
- 230000002996 emotional effect Effects 0.000 claims abstract description 39
- 230000004927 fusion Effects 0.000 claims abstract description 31
- 239000011159 matrix material Substances 0.000 claims abstract description 13
- 238000000605 extraction Methods 0.000 claims abstract description 11
- 230000006870 function Effects 0.000 claims description 39
- 238000012795 verification Methods 0.000 claims description 24
- 230000008451 emotion Effects 0.000 claims description 22
- 238000007781 pre-processing Methods 0.000 claims description 15
- 238000011176 pooling Methods 0.000 claims description 13
- 230000004913 activation Effects 0.000 claims description 10
- 238000012847 principal component analysis method Methods 0.000 claims description 6
- 230000008014 freezing Effects 0.000 claims description 3
- 238000007710 freezing Methods 0.000 claims description 3
- 210000002569 neuron Anatomy 0.000 claims description 3
- 230000007935 neutral effect Effects 0.000 claims 1
- 238000004140 cleaning Methods 0.000 abstract 1
- 238000011160 research Methods 0.000 description 5
- 238000004458 analytical method Methods 0.000 description 2
- 238000003058 natural language processing Methods 0.000 description 2
- 239000013598 vector Substances 0.000 description 2
- 230000009286 beneficial effect Effects 0.000 description 1
- 238000004364 calculation method Methods 0.000 description 1
- 230000007547 defect Effects 0.000 description 1
- 238000011161 development Methods 0.000 description 1
- 238000005516 engineering process Methods 0.000 description 1
- 238000009472 formulation Methods 0.000 description 1
- 238000011835 investigation Methods 0.000 description 1
- 239000000203 mixture Substances 0.000 description 1
- 238000012986 modification Methods 0.000 description 1
- 230000004048 modification Effects 0.000 description 1
- 239000000126 substance Substances 0.000 description 1
- 238000006467 substitution reaction Methods 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F40/00—Handling natural language data
- G06F40/20—Natural language analysis
- G06F40/279—Recognition of textual entities
- G06F40/289—Phrasal analysis, e.g. finite state techniques or chunking
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F16/00—Information retrieval; Database structures therefor; File system structures therefor
- G06F16/30—Information retrieval; Database structures therefor; File system structures therefor of unstructured textual data
- G06F16/33—Querying
- G06F16/3331—Query processing
- G06F16/334—Query execution
- G06F16/3344—Query execution using natural language analysis
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F16/00—Information retrieval; Database structures therefor; File system structures therefor
- G06F16/30—Information retrieval; Database structures therefor; File system structures therefor of unstructured textual data
- G06F16/35—Clustering; Classification
- G06F16/353—Clustering; Classification into predefined classes
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/21—Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
- G06F18/214—Generating training patterns; Bootstrap methods, e.g. bagging or boosting
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/25—Fusion techniques
- G06F18/253—Fusion techniques of extracted features
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
- G06N3/045—Combinations of networks
Abstract
The invention provides a natural language emotion recognition method based on deep learning, which comprises the steps of training a text feature extraction model doc2vec, a picture feature extraction model CNNs and an emotion recognition model CNN according to mass public sentiment data; collecting new public sentiment information, cleaning and classifying public sentiment data, inputting text information into a doc2vec model, and extracting text characteristics; outputting the picture data to a CNNs model, and extracting picture emotional characteristics; performing feature layer fusion on the text and picture emotional features extracted in the first two steps by adopting a feature layer fusion algorithm based on a kernel matrix; and inputting the obtained fusion characteristics into an emotion recognition model and outputting an emotion recognition result.
Description
Technical Field
The invention relates to the technical field of natural language processing, in particular to a natural language emotion recognition method based on deep learning.
Background
In recent years, the internet industry in China is rapidly developed, 45 th reports of a China Internet information center (CNNIC) show that the scale of the netizens in China is 9.04 hundred million and the popularity of the internet reaches 64.5 percent as long as 3 months in 2020, and the scale of users of most network applications is greatly increased. As an open platform, the internet public opinion environment is mixed with the dragon, wherein suggestions and opinions beneficial to social development can provide convenience for the public, and positive energy is brought to people, while partial unfamiliar and distorted public opinions can bring serious psychological panic and safety threat to people and countries if timely and effective public opinion guidance and supervision cannot be carried out.
The network emotion recognition refers to analyzing viewpoints and emotions published by netizens in the internet by a certain method, and dividing the viewpoints and the emotions into different categories according to different standards. Emotion recognition belongs to one of important research contents in the field of natural language processing, and provides essential technical support for government opinion investigation, merchant marketing strategy formulation, internet public opinion guidance and supervision and the like. At present, algorithms and technologies are provided by the academic community aiming at network emotion recognition, most of the researches adopt calculation means, the recognition precision is good when the linguistic data is rich, the emotion tendency is clear and the distribution is uniform, and the network emotion recognition, research and judgment and guidance are well achieved. However, in the face of massive and complex public opinion data in reality, the traditional method sometimes has the problems of unstable performance, low precision and the like, and a new technical method is urgently needed to be sought to break through the existing dilemma.
However, most of the existing network emotion recognition algorithms are based on a single mode of text, and the research of combining text and picture multi-mode features for emotion recognition is not widely concerned by scholars in the field at home and abroad. With the arrival of a new media age, netizens tend to express their emotions in the form of text and images, even sometimes, the same text and different images express completely opposite meanings, so that it is difficult to identify their true emotions sometimes by ignoring the images only through the text, and most of the existing emotion identification algorithms based on the single mode of text adopt word2vec models, neglecting the influence of the arrangement order among words on sentence or text emotion analysis, and easily omitting text hidden information.
In order to solve the above problems, people are always seeking an ideal technical solution.
Disclosure of Invention
Aiming at the defects of the prior art, the invention provides a natural language emotion recognition model training method based on deep learning and a natural language emotion recognition method based on deep learning, so as to solve the problem of inaccurate emotion recognition caused by neglected picture emotion and insufficient superiority of the conventional algorithm model, and effectively improve the emotion recognition accuracy aiming at image-text multimodal data.
In order to achieve the purpose, the invention adopts the technical scheme that: a natural language emotion recognition model training method based on deep learning comprises the following steps:
s1, acquiring mass image-text public sentiment data, and selecting one of the public sentiment data containing a plurality of pictures with the most emotion representative picture; preprocessing the image-text public sentiment data, including removing stop words, segmenting the text, removing meaningless symbols in the text and converting emoticons into corresponding emotional words;
s2, dividing the graph public sentiment data into a training set and a verification set, and dividing the public sentiment data in the training set and the verification set into text data and picture data respectively;
s3, establishing a doc2vec model, wherein a training algorithm adopted by the doc2vec model is a Skip-Gram model, the window length window is set to be 5, the text feature dimension size is set to be 100, the minimum occurrence frequency min _ count of a training word is set to be 3, and the maximum iteration frequency iter in training is set to be 5000;
preprocessing the text data in the training set and the verification set, training a doc2vec model by using the text data in the training set, and verifying the doc2vec model by using the text data in the verification set; outputting a text feature extraction result in each training and verification;
establishing a CNNs model, selecting a VGG-16 model pre-trained by ImageNet as a reference model, modifying the final layer of full-connection output, and replacing the previous full-connection layer; the modified model comprises 13 convolutional layers, 5 pooling layers and 2 full-connection layers; the weights of the last two fully-connected layers are initialized randomly, weights in a pre-trained reference model are used by the other layers, and the number of neurons is set to be 3; freezing the network layer before the last CNN layer without participating in model training, and only keeping the last CNN layer and the network layer after the last CNN layer to participate in training;
preprocessing the picture data in the training set and the verification set, and zooming the picture data by adopting a cv2.resize () function to form picture data matched with the CNNs model; utilizing picture data in a training set to train the CNNs model, and utilizing picture data in a verification set to verify the CNNs model; outputting a picture emotion feature extraction result in each training and verification;
s4, respectively selecting Gaussian kernel functions as kernel functions to obtain kernel matrixes of the text emotional characteristics and the picture emotional characteristics, and performing weighted fusion on the kernel matrixes, wherein the weighting coefficients of the kernel matrixes of the text emotional characteristics and the picture emotional characteristics are 0.75 and 0.25 respectively; selecting a principal component analysis method to perform dimensionality reduction on the fusion feature matrix to obtain fusion features;
and S5, establishing a CNN network, inputting the obtained fusion characteristics into the CNN network for parameter training, and obtaining an emotion recognition model.
The invention also provides a natural language emotion recognition method based on deep learning, which comprises the following steps:
setting a theme target according to the information requirement of a user, collecting mass and complex public opinion data related to the theme target in the Xinlang microblog by utilizing a social media network crawler, and selecting one picture with the most emotion representative from the microblog containing a plurality of pictures;
pre-processing public sentiment data, including removing stop words, segmenting words of a text, removing meaningless symbols in the text, and converting emoticons into corresponding emotional words;
dividing public opinion data into text data and picture data, and respectively preprocessing the text data and the picture data;
inputting the preprocessed text data into a trained doc2vec model for text feature extraction; inputting the preprocessed picture data into a trained CNNs model, and extracting picture emotional characteristics;
respectively selecting Gaussian kernel functions as kernel functions to obtain kernel matrixes of the text emotional characteristics and the picture emotional characteristics, and performing weighted fusion on the kernel matrixes, wherein the weighting coefficients of the kernel matrixes of the text emotional characteristics and the picture emotional characteristics are 0.75 and 0.25 respectively; reducing the dimension of the fusion feature matrix by adopting a principal component analysis method to obtain fusion features;
inputting the obtained fusion characteristics into a trained emotion recognition model, and outputting an emotion recognition result;
wherein the doc2vec model, the CNNs model and the emotion recognition model are trained by adopting the model training method.
Compared with the prior art, the method has outstanding substantive characteristics and remarkable progress, and particularly, the method adopts doc2vec program to extract text features, so that the text feature vector expression is more accurate due to the consideration of the influence of the arrangement sequence among words on sentence or text emotion analysis, and the hidden information of the text is better mined; the picture emotion characteristics are extracted by adopting a CNNs model, a Gaussian function is selected as a kernel function to obtain a kernel matrix, the obtained kernel matrices of two modes are subjected to weighted fusion to obtain fusion characteristics, and then the CNN network is adopted to finish emotion recognition so as to improve the network emotion recognition accuracy, wherein the algorithm adopted by each submodule has higher operability and superiority, so that the method has high practicability in network public opinion recognition and supervision application.
Drawings
Fig. 1 is a schematic flow chart of the identification method of the present invention.
Detailed Description
The technical solution of the present invention is further described in detail by the following embodiments.
The invention provides a natural language emotion recognition model training method based on deep learning, which comprises the following steps:
s1, acquiring mass image-text public sentiment data, and selecting one of the public sentiment data containing a plurality of pictures with the most emotion representative picture; and preprocessing the image-text public sentiment data, including removing stop words, segmenting the text, removing meaningless symbols in the text and converting the emoticons into corresponding emotional words.
And S2, dividing the picture public opinion data into a training set and a verification set, and dividing the public opinion data in the training set and the verification set into text data and picture data respectively.
S3, a doc2vec model is established, a training algorithm adopted by the doc2vec model is a Skip-Gram model, the window length window is set to be 5, the text feature dimension size is set to be 100, the minimum occurrence frequency min _ count of a training word is set to be 3, and the maximum iteration frequency iter in training is set to be 5000.
Preprocessing the text data in the training set and the verification set, training a doc2vec model by using the text data in the training set, and verifying the doc2vec model by using the text data in the verification set; and outputting a text feature extraction result in each training and verification.
Establishing a CNNs model, selecting a VGG-16 model pre-trained by ImageNet as a reference model, modifying the final layer of full-connection output, and replacing the previous full-connection layer; the modified model comprises 13 convolutional layers, 5 pooling layers and 2 full-connection layers; the weights of the last two fully-connected layers are initialized randomly, weights in a pre-trained reference model are used by the other layers, and the number of neurons is set to be 3; and (4) freezing the network layer before the last layer of CNN, not participating in model training, and only keeping the last layer of CNN and the network layer after the last layer of CNN to participate in training.
Specifically, the CNNs model has the following structure:
the first layer and the second layer are two layers of same convolution layers, the input picture pixel is 224 × 3, the image is a three-channel RGB image, the convolution filter pixel is 3 × 3, the step length is 1, the filter number is 64, and the activation function is a ReLU function;
the third layer is a pooling layer, the size of a convolution kernel is 2, the step length is 2, and the filling size is 0;
the fourth layer and the fifth layer are two same convolution layers, the convolution filter pixels are 3 x 3, the step length is 1, the filter number is 64, and the activation function is a ReLU function;
the sixth layer is a pooling layer, the size of a convolution kernel is 2, the step length is 2, and the filling size is 0;
the seventh layer to the ninth layer are three same convolution layers, the convolution filter pixels are 3 x 3, the step length is 1, the filter number is 64, and the activation function is a ReLU function;
the tenth layer is a pooling layer, the size of a convolution kernel is 2, the step length is 2, and the filling size is 0;
the eleventh layer to the thirteenth layer are three same convolution layers, the convolution filter pixels are 3 x 3, the step length is 1, the filter number is 64, and the activation function is a ReLU function;
the fourteenth layer is a pooling layer, the size of the convolution kernel is 2, the step length is 2, and the filling size is 0;
the sixteenth layer to the seventeenth layer are three same convolution layers, the convolution filter pixels are 3 x 3, the step length is 1, the filter number is 64, and the activation function is a ReLU function;
the eighteenth layer is a pooling layer, the convolution kernel size is 2, the step size is 2, the filling size is 0, and an image with 7 × 512 pixels is output;
the nineteenth layer and the twentieth layer are the same full connection layer, the input is the output of the previous layer, and the picture emotional characteristics are output;
the last layer is a Softmax classification layer for generating classification results.
Preprocessing the picture data in the training set and the verification set, and zooming the picture data by adopting a cv2.resize () function to form picture data matched with the CNNs model; utilizing picture data in a training set to train the CNNs model, and utilizing picture data in a verification set to verify the CNNs model; and outputting a picture emotion feature extraction result in each training and verification.
S4, respectively selecting Gaussian kernel functions as kernel functions to obtain kernel matrixes of the text emotional characteristics and the picture emotional characteristics, and performing weighted fusion on the kernel matrixes; selecting a principal component analysis method to perform dimensionality reduction on the fusion feature matrix to obtain fusion features; and inputting the obtained fusion characteristics into a CNN network for parameter training to obtain an emotion recognition model.
the kernel matrix weighted addition method is as follows:
wherein the content of the first and second substances,、a kernel matrix for text data and picture data respectively,a weighting coefficient for the text emotional characteristics,Is the weighting coefficient of the kernel matrix of the picture emotion characteristics,in practical application, the weighting coefficient needs to be dynamically adjusted according to different databases to obtain the optimal result.
In this embodiment, the graphic public opinion data is from the internet public opinion event in the Xinlang microblog, and for the microblog containing a plurality of pictures, only one of the microblogs with the most emotional representativeness is selected, and 2000 groups of texts and pictures are collected, and in addition, 2691561 pieces of training word vectors are obtained by using the microblog data captured in the previous research in recent years and cleaned.
Based on the image-text public opinion data, after text emotional characteristics and picture emotional characteristics are extracted respectively, a Gaussian function is selected as a kernel function to obtain a kernel matrix, the obtained kernel matrices of two modes are subjected to weighted fusion, different weights are set, the weight occupied by the text characteristics is set as the weight occupied by the picture characteristics, the result obtained under different weight settings is shown in the following table, and as can be seen from table 1, when the weight occupied by the text data is 0.75, and the weight occupied by the picture is 0.25, the recognition rate is highest and is 93.1%.
0.85 | 0.75 | 0.65 | 0.5 | 0.35 | 0.25 | 0.15 | |
0.15 | 0.25 | 0.35 | 0.5 | 0.65 | 0.75 | 0.85 | |
Average recognition rate (%) | 90.3% | 93.1% | 90.6% | 83.5% | 80.1% | 60.2% | 55.8% |
TABLE 1 recognition results under different text and picture feature weights
Therefore, in this embodiment, the weighting coefficients of the kernel matrices of the text emotion characteristics and the picture emotion characteristics are respectively selected to be 0.75 and 0.25.
S5, establishing a CNN network, inputting the obtained fusion characteristics into the CNN network for parameter training, and obtaining an emotion recognition model;
it should be noted that, in the CNN network training learning rate η, a strategy that η exponentially decays with the number of iterations is adopted, and the strategy is as follows:
wherein T is the number of iterations, T is the step length, nL is not more thanIs the largest integer of (a).
As shown in fig. 1, the present invention further provides a natural language emotion recognition method based on deep learning, including:
setting a theme target according to the information requirement of a user, collecting mass and complex public opinion data related to the theme target in the Xinlang microblog by utilizing a social media network crawler, and selecting one picture with the most emotion representative from the microblog containing a plurality of pictures;
pre-processing public sentiment data, including removing stop words, segmenting words of a text, removing meaningless symbols in the text, and converting emoticons into corresponding emotional words;
dividing public opinion data into text data and picture data, and respectively preprocessing the text data and the picture data;
inputting the preprocessed text data into a trained doc2vec model for text feature extraction; inputting the preprocessed picture data into a trained CNNs model, and extracting picture emotional characteristics;
respectively selecting Gaussian kernel functions as kernel functions to obtain kernel matrixes of the text emotional characteristics and the picture emotional characteristics, and performing weighted fusion on the kernel matrixes, wherein the weighting coefficients of the kernel matrixes of the text emotional characteristics and the picture emotional characteristics are 0.75 and 0.25 respectively; reducing the dimension of the fusion feature matrix by adopting a principal component analysis method to obtain fusion features;
inputting the obtained fusion characteristics into a trained emotion recognition model, and outputting an emotion recognition result;
wherein the doc2vec model, the CNNs model and the emotion recognition model are trained by adopting the model training method.
Finally, it should be noted that the above examples are only used to illustrate the technical solutions of the present invention and not to limit the same; although the present invention has been described in detail with reference to preferred embodiments, those skilled in the art will understand that: modifications to the specific embodiments of the invention or equivalent substitutions for parts of the technical features may be made; without departing from the spirit of the present invention, it is intended to cover all aspects of the invention as defined by the appended claims.
Claims (3)
1. A natural language emotion recognition model training method based on deep learning is characterized by comprising the following steps:
s1, acquiring mass image-text public sentiment data, and selecting one of the public sentiment data containing a plurality of pictures with the most emotion representative picture; preprocessing the image-text public sentiment data, including removing stop words, segmenting the text, removing meaningless symbols in the text and converting emoticons into corresponding emotional words;
s2, dividing the graph public sentiment data into a training set and a verification set, and dividing the public sentiment data in the training set and the verification set into text data and picture data respectively;
s3, establishing a doc2vec model, wherein a training algorithm adopted by the doc2vec model is a Skip-Gram model, the window length window is set to be 5, the text feature dimension size is set to be 100, the minimum occurrence frequency min _ count of a training word is set to be 3, and the maximum iteration frequency iter in training is set to be 5000;
preprocessing the text data in the training set and the verification set, training a doc2vec model by using the text data in the training set, and verifying the doc2vec model by using the text data in the verification set; outputting a text feature extraction result in each training and verification;
establishing a CNNs model, selecting a VGG-16 model pre-trained by ImageNet as a reference model, modifying the final layer of full-connection output, and replacing the previous full-connection layer; the modified model comprises 13 convolutional layers, 5 pooling layers and 2 full-connection layers; the weights of the last two fully-connected layers are initialized randomly, weights in a pre-trained reference model are used by the other layers, and the number of neurons is set to be 3; freezing the network layer before the last CNN layer without participating in model training, and only keeping the last CNN layer and the network layer after the last CNN layer to participate in training;
preprocessing the picture data in the training set and the verification set, and zooming the picture data by adopting a cv2.resize () function to form picture data matched with the CNNs model; utilizing picture data in a training set to train the CNNs model, and utilizing picture data in a verification set to verify the CNNs model; outputting a picture emotion feature extraction result in each training and verification;
s4, respectively selecting Gaussian kernel functions as kernel functions to obtain kernel matrixes of the text emotional characteristics and the picture emotional characteristics, and performing weighted fusion on the kernel matrixes, wherein the weighting coefficients of the kernel matrixes of the text emotional characteristics and the picture emotional characteristics are 0.75 and 0.25 respectively; selecting a principal component analysis method to perform dimensionality reduction on the fusion feature matrix to obtain fusion features;
and S5, establishing a CNN network, inputting the obtained fusion characteristics into the CNN network for parameter training, and obtaining an emotion recognition model.
2. The method for training the deep learning-based natural language emotion recognition model according to claim 1, wherein the CNNs model has the following structure:
the first layer and the second layer are two layers of same convolution layers, the input picture pixel is 224 × 3, the image is a three-channel RGB image, the convolution filter pixel is 3 × 3, the step length is 1, the filter number is 64, and the activation function is a ReLU function;
the third layer is a pooling layer, the size of a convolution kernel is 2, the step length is 2, and the filling size is 0;
the fourth layer and the fifth layer are two same convolution layers, the convolution filter pixels are 3 x 3, the step length is 1, the filter number is 64, and the activation function is a ReLU function;
the sixth layer is a pooling layer, the size of a convolution kernel is 2, the step length is 2, and the filling size is 0;
the seventh layer to the ninth layer are three same convolution layers, the convolution filter pixels are 3 x 3, the step length is 1, the filter number is 64, and the activation function is a ReLU function;
the tenth layer is a pooling layer, the size of a convolution kernel is 2, the step length is 2, and the filling size is 0;
the eleventh layer to the thirteenth layer are three same convolution layers, the convolution filter pixels are 3 x 3, the step length is 1, the filter number is 64, and the activation function is a ReLU function;
the fourteenth layer is a pooling layer, the size of the convolution kernel is 2, the step length is 2, and the filling size is 0;
the sixteenth layer to the seventeenth layer are three same convolution layers, the convolution filter pixels are 3 x 3, the step length is 1, the filter number is 64, and the activation function is a ReLU function;
the eighteenth layer is a pooling layer, the convolution kernel size is 2, the step size is 2, the filling size is 0, and an image with 7 × 512 pixels is output;
the nineteenth layer and the twentieth layer are the same full connection layer, the input is the output of the previous layer, and the picture emotional characteristics are output;
the last layer is a Softmax classification layer for generating classification results.
3. A natural language emotion recognition method based on deep learning is characterized by comprising the following steps:
setting a theme target according to the information requirement of a user, collecting mass and complex public opinion data related to the theme target in the Xinlang microblog by utilizing a social media network crawler, and selecting one picture with the most emotion representative from the microblog containing a plurality of pictures;
pre-processing public sentiment data, including removing stop words, segmenting words of a text, removing meaningless symbols in the text, and converting emoticons into corresponding emotional words;
dividing public opinion data into text data and picture data, and respectively preprocessing the text data and the picture data;
inputting the preprocessed text data into a trained doc2vec model for text feature extraction; inputting the preprocessed picture data into a trained CNNs model, and extracting picture emotional characteristics;
respectively selecting Gaussian kernel functions as kernel functions to obtain kernel matrixes of the text emotional characteristics and the picture emotional characteristics, and performing weighted fusion on the kernel matrixes, wherein the weighting coefficients of the kernel matrixes of the text emotional characteristics and the picture emotional characteristics are 0.75 and 0.25 respectively; reducing the dimension of the fusion feature matrix by adopting a principal component analysis method to obtain fusion features;
inputting the obtained fusion characteristics into a trained emotion recognition model, and outputting emotion recognition results which are divided into positive, negative and neutral classes;
wherein the doc2vec model, the CNNs model and the emotion recognition model are trained by the training method of claim 1 or 2.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202010613189.7A CN111813894A (en) | 2020-06-30 | 2020-06-30 | Natural language emotion recognition method based on deep learning |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202010613189.7A CN111813894A (en) | 2020-06-30 | 2020-06-30 | Natural language emotion recognition method based on deep learning |
Publications (1)
Publication Number | Publication Date |
---|---|
CN111813894A true CN111813894A (en) | 2020-10-23 |
Family
ID=72856383
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN202010613189.7A Pending CN111813894A (en) | 2020-06-30 | 2020-06-30 | Natural language emotion recognition method based on deep learning |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN111813894A (en) |
Cited By (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN112417097A (en) * | 2020-11-19 | 2021-02-26 | 中国电子科技集团公司电子科学研究院 | Multi-modal data feature extraction and association method for public opinion analysis |
CN112488214A (en) * | 2020-12-02 | 2021-03-12 | 浙江大华技术股份有限公司 | Image emotion analysis method and related device |
CN113901171A (en) * | 2021-09-06 | 2022-01-07 | 特赞(上海)信息科技有限公司 | Semantic emotion analysis method and device |
Citations (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN105740349A (en) * | 2016-01-25 | 2016-07-06 | 重庆邮电大学 | Sentiment classification method capable of combining Doc2vce with convolutional neural network |
CN106250855A (en) * | 2016-08-02 | 2016-12-21 | 南京邮电大学 | A kind of multi-modal emotion identification method based on Multiple Kernel Learning |
-
2020
- 2020-06-30 CN CN202010613189.7A patent/CN111813894A/en active Pending
Patent Citations (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN105740349A (en) * | 2016-01-25 | 2016-07-06 | 重庆邮电大学 | Sentiment classification method capable of combining Doc2vce with convolutional neural network |
CN106250855A (en) * | 2016-08-02 | 2016-12-21 | 南京邮电大学 | A kind of multi-modal emotion identification method based on Multiple Kernel Learning |
Non-Patent Citations (1)
Title |
---|
范涛等: "基于深度学习的多模态融合网民情感识别研究", 《信息资源管理学报》 * |
Cited By (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN112417097A (en) * | 2020-11-19 | 2021-02-26 | 中国电子科技集团公司电子科学研究院 | Multi-modal data feature extraction and association method for public opinion analysis |
CN112488214A (en) * | 2020-12-02 | 2021-03-12 | 浙江大华技术股份有限公司 | Image emotion analysis method and related device |
WO2022116771A1 (en) * | 2020-12-02 | 2022-06-09 | Zhejiang Dahua Technology Co., Ltd. | Method for analyzing emotion shown in image and related devices |
CN113901171A (en) * | 2021-09-06 | 2022-01-07 | 特赞(上海)信息科技有限公司 | Semantic emotion analysis method and device |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN113254599B (en) | Multi-label microblog text classification method based on semi-supervised learning | |
Das et al. | Sign language recognition using deep learning on custom processed static gesture images | |
CN110309839B (en) | A kind of method and device of iamge description | |
CN110287320A (en) | A kind of deep learning of combination attention mechanism is classified sentiment analysis model more | |
CN111813894A (en) | Natural language emotion recognition method based on deep learning | |
CN110362819B (en) | Text emotion analysis method based on convolutional neural network | |
CN109977416A (en) | A kind of multi-level natural language anti-spam text method and system | |
Banerjee et al. | Comparison of pretrained embeddings to identify hate speech in Indian code-mixed text | |
CN109815485B (en) | Method and device for identifying emotion polarity of microblog short text and storage medium | |
CN111552803A (en) | Text classification method based on graph wavelet network model | |
CN111581966A (en) | Context feature fusion aspect level emotion classification method and device | |
CN111475622A (en) | Text classification method, device, terminal and storage medium | |
CN112052687B (en) | Semantic feature processing method, device and medium based on depth separable convolution | |
CN107357785A (en) | Theme feature word abstracting method and system, feeling polarities determination methods and system | |
Pan et al. | Deep neural network-based classification model for Sentiment Analysis | |
CN106446147A (en) | Emotion analysis method based on structuring features | |
CN110472245A (en) | A kind of multiple labeling emotional intensity prediction technique based on stratification convolutional neural networks | |
CN113220890A (en) | Deep learning method combining news headlines and news long text contents based on pre-training | |
CN107818173B (en) | Vector space model-based Chinese false comment filtering method | |
CN110728144B (en) | Extraction type document automatic summarization method based on context semantic perception | |
CN114462385A (en) | Text segmentation method and device | |
CN114625908A (en) | Text expression package emotion analysis method and system based on multi-channel attention mechanism | |
CN109284765A (en) | The scene image classification method of convolutional neural networks based on negative value feature | |
Jadhav et al. | Content based facial emotion recognition model using machine learning algorithm | |
CN111553424A (en) | CGAN-based image data balancing and classifying method |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
RJ01 | Rejection of invention patent application after publication |
Application publication date: 20201023 |
|
RJ01 | Rejection of invention patent application after publication |