CN113516198B - Cultural resource text classification method based on memory network and graphic neural network - Google Patents

Cultural resource text classification method based on memory network and graphic neural network Download PDF

Info

Publication number
CN113516198B
CN113516198B CN202110864647.9A CN202110864647A CN113516198B CN 113516198 B CN113516198 B CN 113516198B CN 202110864647 A CN202110864647 A CN 202110864647A CN 113516198 B CN113516198 B CN 113516198B
Authority
CN
China
Prior art keywords
text
graph
word
words
model
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN202110864647.9A
Other languages
Chinese (zh)
Other versions
CN113516198A (en
Inventor
王海
王妍
黄帝淞
周腾
吴旭东
曹瑞
郑杰
马于惠
高岭
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
NORTHWEST UNIVERSITY
Original Assignee
NORTHWEST UNIVERSITY
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by NORTHWEST UNIVERSITY filed Critical NORTHWEST UNIVERSITY
Priority to CN202110864647.9A priority Critical patent/CN113516198B/en
Publication of CN113516198A publication Critical patent/CN113516198A/en
Application granted granted Critical
Publication of CN113516198B publication Critical patent/CN113516198B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/24Classification techniques
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/21Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
    • G06F18/214Generating training patterns; Bootstrap methods, e.g. bagging or boosting
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F40/00Handling natural language data
    • G06F40/20Natural language analysis
    • G06F40/205Parsing
    • G06F40/211Syntactic parsing, e.g. based on context-free grammar [CFG] or unification grammars
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F40/00Handling natural language data
    • G06F40/20Natural language analysis
    • G06F40/279Recognition of textual entities
    • G06F40/284Lexical analysis, e.g. tokenisation or collocates
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/044Recurrent networks, e.g. Hopfield networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • G06N3/084Backpropagation, e.g. using gradient descent

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Data Mining & Analysis (AREA)
  • Artificial Intelligence (AREA)
  • General Engineering & Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • General Health & Medical Sciences (AREA)
  • Computational Linguistics (AREA)
  • Health & Medical Sciences (AREA)
  • Evolutionary Computation (AREA)
  • Molecular Biology (AREA)
  • Biomedical Technology (AREA)
  • Biophysics (AREA)
  • Computing Systems (AREA)
  • Mathematical Physics (AREA)
  • Software Systems (AREA)
  • Evolutionary Biology (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • Audiology, Speech & Language Pathology (AREA)
  • Information Retrieval, Db Structures And Fs Structures Therefor (AREA)
  • Machine Translation (AREA)

Abstract

A cultural resource text classification method based on a memory network and a graph neural network mainly comprises three modules, a pre-trained two-way long-short term memory network (BiLSTM) module, a text graph construction module and a graph neural network (GCN) module, wherein firstly, a text data set is pre-trained by utilizing a two-way long-short term memory network model and text and word characteristics containing time sequence information are obtained, secondly, a global text graph consisting of the text and the words is constructed according to the co-occurrence relation among the words and the importance of the words in the text, the node characteristics of the text graph are initialized by adopting the characteristics extracted by the pre-trained two-way long-short term memory network module, then, further characterization learning is carried out on the node characteristics of the global text graph through a two-layer graph convolution neural network, and a final text classification result is obtained, so that the method can be used for improving the classification accuracy of cultural resource texts.

Description

Cultural resource text classification method based on memory network and graphic neural network
Technical Field
The invention belongs to the technical field of natural language processing, and particularly relates to a cultural resource text classification method based on a two-way long-short-term memory network and a graphic neural network (BiLSTM-GCN).
Background
Cultural resources are natural or social resources which can meet the requirements of human culture and provide basis for cultural industry, and can be divided into various types. The culture resources are effectively classified, and the storage, the mining and the reuse of the culture resources can be promoted.
Text classification is a technology for automatically classifying and labeling texts according to certain classification rules and classification standards, and is also a basic task of natural language processing, and is used in many applications such as emotion analysis, data mining, news filtering and the like. Early text classification, which relies on manual labeling and some set rules to classify text data, was extremely inefficient, which was not possible to process hundreds of millions of levels of text data at any time in the current big data age. With the progress of the age and the development of computer science and technology, text classification realizes automatic classification by training and learning text data through a computer. The manual classification workload is greatly reduced, the working efficiency is improved, and the classification effect is even better than that of manual classification.
The text classification method mainly comprises a traditional machine learning method and a deep learning method, the traditional machine learning-based text classification method is used for realizing text classification by manually extracting features and utilizing a shallow classifier, the text classification task is very mature, however, as the effect requirements of people on text classification are higher and higher, the current text classification method mainly uses the deep learning-based method, most of deep learning models usually represent unstructured text data into data which can be understood by a computer, and then training is carried out through a large number of marked data sets, so that important features in the text are extracted, and a final classification result is obtained.
The graph neural network is a deep learning network based on graph data, and compared with the traditional network, the graph neural network can better perform characterization learning on the graph structure data, and is widely applied to tasks such as social networks, recommendation systems, molecular activity prediction and the like. The text in the natural language processing task also contains rich graph structure information, including word co-occurrence, syntactic semantics and text context, and the graph neural network can fully utilize the graph structure information of the text data. Current text classification methods with respect to graphic neural networks rarely take into account time series problems in the text, which greatly limit the effectiveness of text classification.
Disclosure of Invention
In order to overcome the problem that the text classification method based on the graph convolution neural network does not fully consider the time sequence problem contained in the text, the invention aims to provide the cultural resource text classification method based on the memory network and the graph neural network, and the method can introduce the time sequence information in the text in advance so as to achieve a better text classification effect.
In order to achieve the above purpose, the technical scheme adopted by the invention is as follows:
a text classification method based on BiLSTM-GCN comprises a BiLSTM pre-training module, a text diagram construction module and a GCN module, and is characterized by comprising the following steps:
1) BiLSTM pre-training module: pre-training the text by using a bidirectional LSTM and acquiring the text and word characteristics after pre-training;
LSTM is a deep learning network based on modeling of time step sequence, its structural features determine that hidden layer at each moment is related to previous input, compared with RNN, LSTM can avoid wellAvoiding the problem of gradient vanishing due to long-term dependence, assuming a text consisting of n words { w ] 1 ,w 2 ,...,w n Each word is represented by a d-dimensional vector, and is initialized by using a glove word vector trained by Stanford, wherein the initialized word vector is { v } 1 ,v 2 ,...,v n In order to get better word vectors for the model, the word vectors are updated continuously with iteration during training. Each text is modeled using bi-directional LSTM, one LSTM is semantic representation of the text from beginning to end, and the other LSTM is semantic representation of the text from end to end, the input at the current time being related not only to the previous state but also to the following state.
Each word v t Both forward and backward hidden statesAnd->
h t The two hidden states are concatenated together, which fuses the word context information. The specific formula is as follows:
through the previous operation, text and word characteristics containing time series information can be extracted, then the full connection layer is needed to classify samples, and the model is used for hiding the first state h 1 And last hidden state h n As the characteristics of the text, the characteristics are input into a full connection layer, and then an activation function is connected to realize text classification, wherein the specific formula is as follows:
y=softmax(MLP(h 1 ,h 2 ));
2) The text diagram construction module: constructing a large global text graph for the whole text data set, wherein nodes in the text graph are composed of texts and words;
constructing a global text graph for the whole text data set, wherein nodes in the text graph are composed of texts and words, the edge weights between words are expressed based on co-occurrence relations between words, PMI values or cosine similarity between words are adopted, the edge weights between the texts and the words are expressed based on importance of the words in the text, and the edge weights are obtained by adopting TF-IDF values or a keyword extraction algorithm;
3) GCN module: the method comprises the steps of realizing characterization learning of text graph nodes through a two-layer graph convolution neural network, and further realizing final text classification;
the word characteristics v and the text characteristics h which contain time sequence information can be obtained through the step 1), the characteristics are initialization characteristics of the global text graph nodes, and the characteristics are put into the GCN, so that the characteristic information of the text graph can be further extracted. Assuming that the constructed undirected text graph is g, the node number is n, and the word node number is n word Wen Benjie the number of points is n text A represents the adjacency matrix of the text graph g, I N Is an identity matrix, D is a degree matrix of the text graph g, W 0 ,W 1 All are weight matrixes, and X is a node characteristic matrix of the text graph g.
The GCN model adopts a graph convolution neural network of semi-supervised learning, the whole model can train part of node data with a small number of labels, then the rest of nodes without labels are classified, and the specific implementation formula is as follows:
the loss function is the cross entropy loss of all text, defined as follows:
Y D is a text with labels, i.e. training set and validation set in the model, F is the dimension of the final output feature of the model, equal to the number of classifications, y pred Is a label for model prediction.
The beneficial effects of the invention are as follows:
the text and word characteristics are initialized through the pre-trained BiLSTM module, time sequence information contained in the text can be introduced in advance before graph convolution is carried out, the defect that the global text graph cannot contain the time sequence information is overcome, and then the text graph is characterized and learned through a two-layer graph convolution neural network, so that the effect of text classification can be greatly improved.
Drawings
FIG. 1 is a schematic diagram of a method of the present invention.
Detailed Description
The present invention will be further described with reference to the accompanying drawings and examples, but the present invention is not limited to the following examples.
As shown in FIG. 1, the text classification method based on BiLSTM-GCN comprises a BiLSTM pre-training module, a text diagram construction module and a GCN module, and is characterized by comprising the following steps:
1) BiLSTM pre-training module: pre-training the text by using a bidirectional LSTM and acquiring the text and word characteristics after pre-training;
LSTM is a deep learning network based on modeling of time step sequence, its structural features determine that hidden layer at each moment is related to previous input, compared with RNN, LSTM can well avoid gradient vanishing problem caused by long-time dependence, and a text is assumed to be composed of n words { w 1 ,w 2 ,...,w n Each word is represented by a d-dimensional vector, and is initialized by using a glove word vector trained by Stanford, wherein the initialized word vector is { v } 1 ,v 2 ,...,v n In order to get better word vectors for the model, the word vectors are updated continuously with iteration during training. Each text is modeled using bi-directional LSTM, one LSTM is semantic representation of the text from beginning to end, and the other LSTM is semantic representation of the text from end to end, the input at the current time being related not only to the previous state but also to the following state.
Each word v t Both forward and backward hidden statesAnd->
h t The two hidden states are concatenated together, which fuses the word context information. The specific formula is as follows:
through the previous operation, text and word characteristics containing time series information can be extracted, then the full connection layer is needed to classify samples, and the model is used for hiding the first state h 1 And last hidden state h n As the characteristics of the text, the characteristics are input into a full connection layer, and then an activation function is connected to realize text classification, wherein the specific formula is as follows:
y=softmax(MLP(h 1 ,h 2 ));
2) The text diagram construction module: constructing a large global text graph for the whole text data set, wherein nodes in the text graph are composed of texts and words;
constructing a global text graph for the whole text data set, wherein nodes in the text graph are composed of texts and words, the edge weights between words are expressed based on co-occurrence relations between words, PMI values or cosine similarity between words are adopted, the edge weights between the texts and the words are expressed based on importance of the words in the text, and the edge weights are obtained by adopting TF-IDF values or a keyword extraction algorithm;
3) GCN module: the method comprises the steps of realizing characterization learning of text graph nodes through a two-layer graph convolution neural network, and further realizing final text classification;
word features v and text features h containing time sequence information can be obtained through the step 1), the features are initialization features of the global text graph nodes, and the features are put into GCN and can be furtherFeature information of the text graph is extracted. Assuming that the constructed undirected text graph is g, the node number is n, and the word node number is n word Wen Benjie the number of points is n text A represents the adjacency matrix of the text graph g, I N Is an identity matrix, D is a degree matrix of the text graph g, W 0 ,W 1 All are weight matrixes, and X is a node characteristic matrix of the text graph g.
The GCN model adopts a graph convolution neural network of semi-supervised learning, the whole model can train part of node data with a small number of labels, then the rest of nodes without labels are classified, and the specific implementation formula is as follows:
the loss function is the cross entropy loss of all text, defined as follows:
Y D is a text with labels, i.e. training set and validation set in the model, F is the dimension of the final output feature of the model, equal to the number of classifications, y pred Is a label for model prediction.
Experimental comparative analysis process
The experiment of the invention is based on Python3.7 programming language and Pytorch frame, the word vector uses 100-dimensional, 200-dimensional and 300-dimensional Glove word vector of Stanford pre-training, and the initial learning rate of the whole model is 10 -3 L2 weight decay is 10 -4 Dropou after full connection layert is set to 0.5 and the batch size is set to 64. If the validation set does not drop in loss after consecutive 5 generations of training, training is stopped.
This patent was tested on the R8 and R52 datasets to verify the validity of the model and its performance.
R8, R52 are each news classification data sets, wherein each sample contains a piece of news text and a corresponding hot topic label. Wherein, R8 contains 8 news categories, the training set and the test set respectively have 5485 samples and 2189 samples, R52 contains 52 news categories, and the training set and the test set respectively have 6532 samples and 2568 samples. In the R8, R52 data sets, 10% of the training set data were randomly selected as the validation set. The comparative experiments are mainly divided into algorithms based on a traditional deep learning algorithm comprising CNN, LSTM, biLSTM and Fastext and algorithms based on a graph neural network comprising Text-GCN.
The accuracy and the F1 value can be used as evaluation indexes of the performance of the classification model, as shown in table 1, the text classification model based on the GCN is stronger than the traditional deep learning text classification model on several classification data sets, and the text classification model based on the GCN can learn rich graph information in the text due to the fact that graph convolution is adopted for learning text characteristics, so that better classification effect can be achieved. According to the BiLSTM-GCN model, time sequence information of a text is introduced in advance on the basis of GCN, characteristic information of the text is enriched, experiments show that the accuracy of the BiLSTM-GCN model on an R8 data set is improved by 0.51%, and the accuracy of the BiLSTM-GCN model on an R52 data set is improved by 0.7%, and the fact that a better classification effect is achieved by directly using one-hot characteristics after words brought by the BiLSTM model are introduced and the text characteristic information is achieved.
TABLE 1 BiLSTM-GCN vs. different model accuracy
TABLE 2 comparison of BiLSTM-GCN with different model F1 values
The accuracy is only used, so that the effect of the model is not sufficiently evaluated, the recall rate and the accuracy are considered by the F1 value, and the effect of the model provided by the invention can be more sufficiently judged. As can be seen from table 2, the model proposed by the present invention has an improvement of 1.32% over the R8 dataset and 1.84% over the R52 dataset over the baseline GCN model F1, which further indicates that the model proposed by us is more effective in text classification tasks.

Claims (1)

1. A cultural resource text classification method based on a memory network and a graphic neural network is characterized by comprising the following steps:
1) BiLSTM pre-training module: pre-training the text by using a bidirectional LSTM and acquiring the text and word characteristics after pre-training;
assume that a text consists of n words { w } 1 ,w 2 ,…,w n Each word is represented by a d-dimensional vector, and is initialized by using a glove word vector trained by Stanford, wherein the initialized word vector is { v } 1 ,v 2 ,…v n In order to make the model obtain better word vectors, the word vectors are updated continuously along with iteration in the training process; modeling each text by adopting a bidirectional LSTM, wherein one LSTM is used for carrying out semantic representation of the text from the beginning of the sentence to the end of the sentence, and the other LSTM is used for carrying out semantic representation of the text from the end of the sentence to the beginning of the sentence, and the input at the current moment is not only related to the previous state but also related to the following state;
each word v t Both forward and backward hidden statesAnd->h t The two hidden states are spliced together, and the two hidden states fuse the context information of the word, and the specific formula is as follows:
through the previous operation, text and word characteristics containing time series information can be extracted, then the full connection layer is needed to classify samples, and the model is used for hiding the first state h 1 And last hidden state h n As the characteristics of the text, the characteristics are input into a full connection layer, and then an activation function is connected to realize text classification, wherein the specific formula is as follows:
y=soft max(MLP(h 1 ,h n ));
2) The text diagram construction module: constructing a large global text graph for the whole text data set, wherein nodes in the text graph are composed of texts and words;
constructing a global text graph for the whole text data set, wherein nodes in the text graph are composed of texts and words, the edge weights between words are expressed based on co-occurrence relations between words, PMI values or cosine similarity between words are adopted, the edge weights between the texts and the words are expressed based on importance of the words in the text, and the edge weights are obtained by adopting TF-IDF values or a keyword extraction algorithm;
3) GCN module: the method comprises the steps of realizing characterization learning of text graph nodes through a two-layer graph convolution neural network, and further realizing final text classification;
the word characteristics v and text characteristics h containing time sequence information can be obtained through the step 1), the characteristics are initialization characteristics of the global text graph nodes, the characteristics are put into GCN, the characteristic information of the text graph can be further extracted, the constructed undirected text graph is assumed to be g, the node number is N, and the word node number is N word Wen Benjie the number of points is N text A represents the adjacency matrix of the text graph g, I N Is an identity matrix, W 0 ,W 1 The node feature matrices are weight matrices, and X is a node feature matrix of the text graph g;
the GCN model adopts a graph convolution neural network of semi-supervised learning, the whole model can train part of node data with a small number of labels, then classify the rest of nodes without labels, and the specific implementation formula is as follows:
the loss function is the cross entropy loss of all text, defined as follows:
Y D is a text with labels, i.e. training set and validation set in the model, F is the dimension of the final output feature of the model, equal to the number of classifications, y pred Is a label for model prediction.
CN202110864647.9A 2021-07-29 2021-07-29 Cultural resource text classification method based on memory network and graphic neural network Active CN113516198B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202110864647.9A CN113516198B (en) 2021-07-29 2021-07-29 Cultural resource text classification method based on memory network and graphic neural network

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202110864647.9A CN113516198B (en) 2021-07-29 2021-07-29 Cultural resource text classification method based on memory network and graphic neural network

Publications (2)

Publication Number Publication Date
CN113516198A CN113516198A (en) 2021-10-19
CN113516198B true CN113516198B (en) 2023-09-22

Family

ID=78068765

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202110864647.9A Active CN113516198B (en) 2021-07-29 2021-07-29 Cultural resource text classification method based on memory network and graphic neural network

Country Status (1)

Country Link
CN (1) CN113516198B (en)

Families Citing this family (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN117455421B (en) * 2023-12-25 2024-04-16 杭州青塔科技有限公司 Subject classification method and device for scientific research projects, computer equipment and storage medium

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN106021461A (en) * 2016-05-17 2016-10-12 深圳市中润四方信息技术有限公司 Text classification method and text classification system
CN109597891A (en) * 2018-11-26 2019-04-09 重庆邮电大学 Text emotion analysis method based on two-way length Memory Neural Networks in short-term
US10339423B1 (en) * 2017-06-13 2019-07-02 Symantec Corporation Systems and methods for generating training documents used by classification algorithms
WO2021008037A1 (en) * 2019-07-15 2021-01-21 平安科技(深圳)有限公司 A-bilstm neural network-based text classification method, storage medium, and computer device
CN112966503A (en) * 2021-03-22 2021-06-15 山东建筑大学 Aspect level emotion analysis method

Family Cites Families (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US11481418B2 (en) * 2020-01-02 2022-10-25 International Business Machines Corporation Natural question generation via reinforcement learning based graph-to-sequence model

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN106021461A (en) * 2016-05-17 2016-10-12 深圳市中润四方信息技术有限公司 Text classification method and text classification system
US10339423B1 (en) * 2017-06-13 2019-07-02 Symantec Corporation Systems and methods for generating training documents used by classification algorithms
CN109597891A (en) * 2018-11-26 2019-04-09 重庆邮电大学 Text emotion analysis method based on two-way length Memory Neural Networks in short-term
WO2021008037A1 (en) * 2019-07-15 2021-01-21 平安科技(深圳)有限公司 A-bilstm neural network-based text classification method, storage medium, and computer device
CN112966503A (en) * 2021-03-22 2021-06-15 山东建筑大学 Aspect level emotion analysis method

Non-Patent Citations (3)

* Cited by examiner, † Cited by third party
Title
Graph Convolutional Networks for Text Classification;Liang Yao 等;《arXiv》;20181113;第1-9页 *
基于BERT的文本情感分析;刘思琴等;《信息安全研究》;20200305(第03期);第30-37页 *
基于预训练模型的文本分类网络TextCGA;杨玮祺等;《现代计算机》;20200425(第12期);第53-38页 *

Also Published As

Publication number Publication date
CN113516198A (en) 2021-10-19

Similar Documents

Publication Publication Date Title
CN110866117B (en) Short text classification method based on semantic enhancement and multi-level label embedding
CN110413986B (en) Text clustering multi-document automatic summarization method and system for improving word vector model
CN111274394B (en) Method, device and equipment for extracting entity relationship and storage medium
CN106599022B (en) User portrait forming method based on user access data
CN112699247B (en) Knowledge representation learning method based on multi-class cross entropy contrast complement coding
CN110321563B (en) Text emotion analysis method based on hybrid supervision model
CN111078833B (en) Text classification method based on neural network
CN114896388A (en) Hierarchical multi-label text classification method based on mixed attention
CN107688576B (en) Construction and tendency classification method of CNN-SVM model
CN111274790B (en) Chapter-level event embedding method and device based on syntactic dependency graph
CN114461804B (en) Text classification method, classifier and system based on key information and dynamic routing
CN111506732A (en) Text multi-level label classification method
CN113806547B (en) Deep learning multi-label text classification method based on graph model
CN112925904B (en) Lightweight text classification method based on Tucker decomposition
CN113254655B (en) Text classification method, electronic device and computer storage medium
CN114925205B (en) GCN-GRU text classification method based on contrast learning
CN112749274A (en) Chinese text classification method based on attention mechanism and interference word deletion
CN113051887A (en) Method, system and device for extracting announcement information elements
CN116049387A (en) Short text classification method, device and medium based on graph convolution
CN113343690A (en) Text readability automatic evaluation method and device
CN115659947A (en) Multi-item selection answering method and system based on machine reading understanding and text summarization
CN111709225A (en) Event cause and effect relationship judging method and device and computer readable storage medium
CN113516198B (en) Cultural resource text classification method based on memory network and graphic neural network
CN112463982B (en) Relationship extraction method based on explicit and implicit entity constraint
CN113486143A (en) User portrait generation method based on multi-level text representation and model fusion

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant