CN110516070B - Chinese question classification method based on text error correction and neural network - Google Patents

Chinese question classification method based on text error correction and neural network Download PDF

Info

Publication number
CN110516070B
CN110516070B CN201910801515.4A CN201910801515A CN110516070B CN 110516070 B CN110516070 B CN 110516070B CN 201910801515 A CN201910801515 A CN 201910801515A CN 110516070 B CN110516070 B CN 110516070B
Authority
CN
China
Prior art keywords
vector
matrix vector
text data
attention
chinese question
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN201910801515.4A
Other languages
Chinese (zh)
Other versions
CN110516070A (en
Inventor
杨一何
刘晋
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Shanghai Maritime University
Original Assignee
Shanghai Maritime University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Shanghai Maritime University filed Critical Shanghai Maritime University
Priority to CN201910801515.4A priority Critical patent/CN110516070B/en
Publication of CN110516070A publication Critical patent/CN110516070A/en
Application granted granted Critical
Publication of CN110516070B publication Critical patent/CN110516070B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/30Information retrieval; Database structures therefor; File system structures therefor of unstructured textual data
    • G06F16/35Clustering; Classification
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Data Mining & Analysis (AREA)
  • General Engineering & Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Biomedical Technology (AREA)
  • Evolutionary Computation (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Artificial Intelligence (AREA)
  • Databases & Information Systems (AREA)
  • Biophysics (AREA)
  • Computational Linguistics (AREA)
  • Health & Medical Sciences (AREA)
  • General Health & Medical Sciences (AREA)
  • Molecular Biology (AREA)
  • Computing Systems (AREA)
  • Mathematical Physics (AREA)
  • Software Systems (AREA)
  • Machine Translation (AREA)
  • Information Retrieval, Db Structures And Fs Structures Therefor (AREA)

Abstract

The invention discloses a Chinese question classification method based on text error correction and a neural network, which comprises the following steps: acquiring text data of Chinese question sentences; correcting the text data to obtain corrected text data; preprocessing the corrected text data to obtain a Chinese question matrix vector; inputting the Chinese question matrix vector to a bidirectional gating circulation unit layer to obtain a middle semantic matrix vector; obtaining an attention matrix vector according to the attention weight value corresponding to the intermediate semantic matrix vector; inputting the attention moment matrix vector into a convolution neural network layer to obtain a global feature matrix vector; inputting the global characteristic matrix vector into a full-connection layer to obtain probability distribution of each category; and obtaining a Chinese question classification result based on the probability distribution. By applying the embodiment of the invention, the input question is corrected firstly, and then the bidirectional gating circulation unit network model, the attention mechanism and the convolutional neural network model are combined, so that the classification is more accurate.

Description

Chinese question classification method based on text error correction and neural network
Technical Field
The invention relates to the technical field of intelligent information processing and computers, in particular to a Chinese question classification method based on text error correction and a neural network.
Background
With rapid development of science and technology in the internet era and mass data coming in, the returned results need to be manually screened when a search engine is used for searching key words, and time and labor are consumed for users. The intention of the user can be rapidly obtained through the question-answering system, and the most accurate answer can be returned to the user from hundreds of candidate answers.
The Chinese question classification is the first step of a question-answering system and is one of key technologies for realizing accurate answer of the question-answering system, and through the classification of Chinese question, the question-answering system can effectively narrow the answer range and determine the question processing mode, so that the answer of the question-answering system is more accurate and reliable.
At present, the research on Chinese question classification methods mainly includes three types: one is a method based on rule matching and feature extraction, the other is a method based on traditional machine learning, and the last is a method based on deep learning. The method based on rule matching and feature extraction defines a set of rules for the features of different Chinese question sentences, and realizes the classification of the Chinese question sentences by analyzing the matching degree of the Chinese question sentences and the rules. There are many Chinese question classification methods based on machine learning, and the common methods include naive Bayes classification, support vector machine classification and the like, but still need to actively extract features, and still have certain subjectivity for Chinese question classification. The Chinese question classification method based on the deep learning also comprises a plurality of methods including convolutional neural network classification, cyclic neural network classification and the like, and compared with the traditional method, the Chinese question classification based on the deep learning method has higher accuracy. The current research and application proves that the upper and lower semantic information of Chinese question can be better learned based on a cyclic neural network such as long-short term memory, a bidirectional gating cyclic unit and the like, but the current research and application is not good at extracting local characteristics such as key information and the like. The convolutional neural network based on the convolutional neural network can learn local features in the sentence better, key information in the sentence is extracted, and position information of words can be omitted. At present, Chinese question sentences are classified in various types, the classification effect of a single classification method still cannot completely meet the requirement in practice, and few researches at present can fully utilize the advantages of a cyclic neural network and a convolutional neural network.
In addition, most of the current methods for classifying Chinese question sentences by using a deep learning neural network almost do not consider the situations of language diseases, multiple characters, different characters and the like of input question sentences, and all the situations are used as the input of a model for prediction or training, so that the trained model has larger deviation for similar question sentences, and the prediction is wrong.
Therefore, the problem of low classification accuracy caused by language diseases, different characters, multiple characters and the like of input question and the inherent defect of single classification method exists in the traditional Chinese question classification.
Disclosure of Invention
The invention aims to provide a Chinese question classification method based on text error correction and a neural network, and aims to solve the problem that the classification accuracy is not high enough due to the fact that the input question has language diseases, different characters and multiple characters and the existing classification method is single.
In order to achieve the above object, the present invention provides a method for classifying chinese question sentences based on text error correction and neural network, the method comprising:
acquiring text data of Chinese question sentences;
correcting the text data to obtain corrected text data;
preprocessing the error-corrected text data to obtain a Chinese question matrix vector;
inputting the Chinese question matrix vector to a bidirectional gating circulation unit layer to obtain a middle semantic matrix vector;
obtaining an attention matrix vector according to the attention weight corresponding to the intermediate semantic matrix vector;
inputting the attention moment matrix vector to a convolutional neural network layer to obtain a global feature matrix vector;
inputting the global feature matrix vector to a full-connection layer to obtain probability distribution of each category;
and obtaining a Chinese question classification result based on the probability distribution.
In one implementation manner of the present invention, the step of correcting the error of the text data to obtain the text data after error correction includes:
detecting error characters in the text data of the Chinese question to obtain a suspected error character candidate set;
traversing the obtained suspected error character candidate set, replacing error characters by using a similar pronunciation dictionary, and obtaining an error-corrected text data candidate set;
and calculating sentence confusion of the obtained text data candidate set after error correction through a language model, comparing and sequencing calculation results, and obtaining error-corrected text data.
In one implementation, the step of preprocessing the error-corrected text data to obtain a chinese question matrix vector includes:
removing target characters in the error-corrected text data to obtain preprocessed text data, wherein the target characters comprise preset characters and preset symbols;
and performing word segmentation processing on the obtained preprocessed text data to obtain a word segmentation processing result.
And vectorizing the word segmentation processing result by using a word vector tool to obtain a Chinese question matrix vector.
In one implementation, the step of inputting the matrix vector of the chinese question sentence into the bidirectional gated cyclic unit layer to obtain the intermediate semantic matrix vector includes:
inputting the Chinese question matrix vector into a bidirectional gating circulation unit layer, wherein the bidirectional gating circulation unit layer comprises a plurality of gating circulation units, and a single word vector is input into each gating circulation unit layer according to the sequence of words in the Chinese question;
and splicing the output of each gating circulation unit to generate an intermediate semantic matrix vector.
In one implementation, the step of inputting the attention moment matrix vector to a convolutional neural network layer to obtain a global feature matrix vector includes:
inputting the attention moment array vector into a convolutional neural network layer to enable the convolutional neural network layer to calculate convolutional kernels with different lengths, and obtaining a convolution result;
performing maximum pooling processing on the convolution result, and extracting local features;
and splicing the extracted local features to obtain a global feature matrix vector.
In one implementation, the step of inputting the global feature matrix vector to a full connection layer to obtain probability distributions of respective categories includes:
inputting the global feature matrix vector to a full connection layer to obtain an output vector;
and converting the output vector into probability distribution of corresponding categories according to the normalized exponential function.
Preferably, the step of obtaining the attention matrix vector according to the attention weight corresponding to the intermediate semantic matrix vector includes:
calculating an attention value by adopting a normalization exponential function according to the output of the gating circulation unit at the current moment and the attention weight parameter;
calculating the attention vector at the current moment according to the current moment and the total length of the time sequence;
and splicing the attention vectors at different moments to obtain an attention moment array vector.
By applying the Chinese question classification method based on text error correction and the neural network, the text error correction is carried out on the input question before the input question is classified in use, and after the text error correction, a bidirectional gated cyclic unit network model, a self-attention model and a convolutional neural network model are combined. The problem that the classification accuracy is not high enough due to the inherent defects of word diseases, different characters and multiple characters existing in the input question and the single classification method in the prior art is solved, so that the classification is more accurate.
Drawings
FIG. 1 is a schematic flow chart of an embodiment of the present invention.
Fig. 2 is a schematic structural diagram according to an embodiment of the present invention.
Detailed Description
The embodiments of the present invention are described below with reference to specific embodiments, and other advantages and effects of the present invention will be easily understood by those skilled in the art from the disclosure of the present specification. The invention is capable of other and different embodiments and of being practiced or of being carried out in various ways, and its several details are capable of modification in various respects, all without departing from the spirit and scope of the present invention.
Please refer to fig. 1-2. It should be noted that the drawings provided in the present embodiment are only for illustrating the basic idea of the present invention, and the components related to the present invention are only shown in the drawings rather than drawn according to the number, shape and size of the components in actual implementation, and the type, quantity and proportion of the components in actual implementation may be changed freely, and the layout of the components may be more complicated.
The invention provides a Chinese question classification method based on text error correction and neural network, as shown in fig. 1-2, comprising:
s110, Chinese question text data is obtained.
It should be noted that the obtained text data of the chinese question may be a search question input by the user in a search engine, a question input in a question-and-answer system, or a text type chinese question directly input by the user, which is not limited herein.
And S120, correcting the text data to obtain corrected text data.
Detecting the characters in the text data by using the language model confusion degree, calculating the likelihood probability value of the characters, and adding a suspected error character candidate set if the likelihood probability value is lower than the sentence text average value;
traversing the obtained suspected error character candidate set, replacing error characters by using a similar pronunciation dictionary, and obtaining an error-corrected text data candidate set; and calculating sentence confusion of the obtained text data candidate set after error correction through a language model, comparing and sequencing calculation results, and obtaining error-corrected text data. The language model may be an N-gram language model, a maximum entropy model, a neural network model, or the like, and is not limited herein.
S130, preprocessing the corrected text data to obtain a Chinese question matrix vector.
It can be understood that before the word segmentation processing is performed on the text data of the Chinese question, the special characters and the symbol punctuations in the Chinese question are removed first, and only the characters are reserved.
Then, the word segmentation processing is carried out to obtain a word segmentation result sequence w 1 ,w 2 ,...,w n-1 ,w n Where n is the number of words in the Chinese question, w i For a word in a Chinese question, each w i Are all words in the domain dictionary.
It can be understood that the word vector model trained in advance is adopted to vectorize the Chinese question after error correction and word segmentation to obtain a two-dimensional matrix vector, namely the word vector matrix of the Chinese question
S 1 ,S 2 ,...,S n-1 ,S n . Wherein S is i ∈R d ,S i The method is characterized in that the method is a word vector of each word in a Chinese question, d is the dimension of the word vector, n is the length of the Chinese question, the dimension of the whole Chinese question matrix vector is dXn, and a word vector tool is a Google word vector tool.
S140, inputting the Chinese question matrix vector to a bidirectional gating circulation unit layer to obtain a middle semantic matrix vector;
the word vector matrix S of Chinese question 1 ,S 2 ,...,S n-1 ,S n The word vector of each word corresponds to one gating circulation unit, and the gating circulation unit layer inputs single word vectors in sequence according to the sequence of the words in the Chinese question. The input sequence feature expression of the Chinese question adopts a directed gating circulation unit, namely the directed gating circulation unit and a reverse gating circulation unit are included, the network structure of the reverse gating circulation unit is the same as that of the forward gating circulation unit, and only the sequence of the input sequence is opposite. Therefore, the input Chinese question word vector matrix is input into the forward gated cyclic unit and the backward gated cyclic unit, i.e. the bidirectional gated cyclic unit.
Each gated cyclic unit outputs vector values of specified dimensions, the outputs of each gated cyclic unit are spliced to generate an intermediate semantic matrix vector H containing upper and lower semantic information 1 ,H 2 ,...,H n-1 ,H n In which H is i ∈R k And k is the specified dimension size.
The process of outputting the vector value of the designated dimension by each gated loop unit is as follows:
gating the cyclic unit involves 4 parts of computation, let R t ,Z t
Figure BDA0002182454640000061
And H t Respectively representing the reset vector, the refresh vector, the candidate memory cell at time tAnd outputting, wherein the time t represents the input sequence of the Chinese question word vectors;
calculating a reset vector R t The formula is as follows,
R t =σ(W r S t +U r H t-1 +B r )
wherein, W r And U r Two variables are weight parameters, H t-1 Is the output, S, of the last time-gated loop unit t For the word vector input at the present moment, B r Is a bias parameter, σ is an activation function;
calculating an update vector Z t The formula is as follows:
Z t =σ(W z S t +U z H t-1 +B z )
wherein, W z And U z Two variables are weight parameters, H t-1 Is the output of the last gated circulation cell, S t For the word vector input at the present moment, B z Is a bias parameter, σ is an activation function;
computing candidate memory cells
Figure BDA0002182454640000062
The formula is as follows,
Figure BDA0002182454640000063
wherein, W and U are weight parameters, B is bias parameter, and H t-1 Is the output, S, of the last-moment gated loop unit t For the word vector input at the present moment, R t For the calculated reset vector, tanh is a tangent function;
calculating the output H at the current moment t The formula is as follows:
Figure BDA0002182454640000064
wherein H t-1 Is at the last momentOutput of a gated cyclic unit, Z t Is the calculated update vector and is the update vector,
Figure BDA0002182454640000065
are candidate memory units for computation.
And S150, obtaining an attention matrix vector according to the attention weight value corresponding to the intermediate semantic matrix vector.
Note that the attention value a is calculated ij The formula is as follows:
a ij =softmax(W a2 tanh(W a1 H t ))
wherein, W a2 And W a1 As a weight parameter, H t Gating the output of the cyclic unit at the current moment;
calculating the attention vector C at the current moment t The formula is as follows:
Figure BDA0002182454640000071
wherein, C t For the output of the self-attention mechanism at the present moment, T h Is the total length of the sequence, t is the current time, a ij Is the attention weight value at the current moment;
splicing the attention vectors at different moments to obtain an attention moment array vector
Figure BDA0002182454640000072
And S160, inputting the attention moment matrix vector into a convolutional neural network layer to obtain a global feature matrix vector.
It will be appreciated that the moment matrix vector will be noted
Figure BDA0002182454640000073
As input to the convolutional neural network layer, wherein the convolutional neural network layer uses convolutional kernels of different lengths for calculation, and maximum pooling decimation is performed on the convolutional resultsAnd (4) taking local features, splicing a plurality of pooled results to obtain a global feature vector as the output of the convolutional neural network layer.
The attention moment array vector
Figure BDA0002182454640000074
At each time C t Is a specified dimension k, and the length of the moment array vector is a specified time sequence length T h Therefore, the attention matrix vector can be regarded as a k × T h Each column represents an attention vector for a word.
The method for extracting the input features by using different convolution kernels comprises the following specific steps:
setting n different convolution kernel sizes, n is greater than or equal to 2, wherein the length of the convolution kernel is constant and is the dimension k of the attention vector, the width m can be changed, and m is greater than or equal to 1 and less than or equal to T h
Different convolution kernels simultaneously carry out convolution calculation on input, and the calculation formula of the convolution kernels is as follows:
T i =f(WC i:i+m-1 +b)
where W is the weight information, m is the width of the convolution kernel, b is the bias parameter, C i:i+m-1 The method is the splicing of attention vectors at a plurality of moments, and the formula is as follows:
Figure BDA0002182454640000075
wherein, C i Attention vector at time i;
performing maximum pooling on the convolution result to extract local features, wherein the formula is as follows:
Figure BDA0002182454640000076
wherein max is the calculation of T i Maximum value of (1);
and splicing a plurality of pooled results to obtain a global feature vector as the output of a convolutional neural network layer, wherein the formula is as follows:
Figure BDA0002182454640000081
where n is the number of different convolution kernels and T is the output of the convolutional neural network layer.
And S170, inputting the global feature matrix vector to a full-connection layer to obtain probability distribution of each category.
It should be noted that the global feature vector obtained by calculation is used as the input of the Fully connected layer (composed of a Fully connected layer), and the output vector of the Fully connected layer is calculated;
the formula is as follows:
F=σ(W f T+B f )
wherein, W f Is a weight parameter, B f Is a bias parameter, σ is the activation function, and T is the global feature vector.
And then converting the output vector into probability distribution of corresponding categories according to the normalized exponential function.
It should be noted that, the normalized exponential function is calculated to obtain the probability distribution of all classes of the model (the classification function Softmax function is used for probability classification); calculating a normalized exponential function to obtain the output of the model, wherein the formula is as follows:
Figure BDA0002182454640000082
F j for the j-th dimension, O, of the fully-connected layer output vector i Is the probability of the ith classification result.
S180, obtaining a Chinese question classification result based on the probability distribution.
And processing the global characteristic matrix vector as the input of the full-connection layer, converting the output of the full-connection layer into the probability distribution of corresponding categories by utilizing a normalized exponential function, wherein the category with the maximum probability value is the predicted category, namely the classification result of the Chinese question (Label is adopted to represent the Label of the classification result). Therefore, the index where the maximum output value is located is taken as the category of the Chinese question, and the Chinese question classification result can be obtained.
It can be understood that after error correction, the invention firstly converts the word vector matrix of the Chinese question into the intermediate semantic matrix vector containing context semantic information by passing the correct Chinese question through the bidirectional gating circulation unit layer. And then, effectively strengthening the function of key words in the Chinese question by using a self-attention mechanism, then extracting local features of the attention matrix vector by using a convolutional neural network and pooling operation, and finally outputting a Chinese question classification result through a full connection layer and a normalized exponential function. By combining the bidirectional gating cyclic unit network model, the self-attention model and the convolutional neural network model, the advantages of different models can be fully exerted, and therefore classification is more accurate.
The foregoing embodiments are merely illustrative of the principles and utilities of the present invention and are not intended to limit the invention. Any person skilled in the art can modify or change the above-mentioned embodiments without departing from the spirit and scope of the present invention. Accordingly, it is intended that all equivalent modifications or changes which can be made by those skilled in the art without departing from the spirit and technical spirit of the present invention be covered by the claims of the present invention.

Claims (7)

1. A Chinese question classification method based on text error correction and neural network is characterized by comprising the following steps:
acquiring text data of Chinese question sentences;
correcting the text data to obtain corrected text data;
preprocessing the error-corrected text data to obtain a Chinese question matrix vector;
inputting the Chinese question matrix vector into a bidirectional gating circulation unit layer to obtain a middle semantic matrix vector;
obtaining an attention matrix vector according to the attention weight corresponding to the intermediate semantic matrix vector;
inputting the attention moment matrix vector into a convolutional neural network layer to obtain a global feature matrix vector;
inputting the global feature matrix vector to a full-connection layer to obtain probability distribution of each category;
and obtaining a Chinese question classification result based on the probability distribution.
2. The method for classifying Chinese question sentences based on text error correction and neural network as claimed in claim 1, wherein the step of correcting the text data to obtain the text data after error correction comprises:
detecting error characters in the text data of the Chinese question sentence to obtain a suspected error character candidate set;
traversing the obtained suspected wrong character candidate set, and replacing wrong characters by using a sound-like dictionary to obtain an error-corrected text data candidate set;
and calculating sentence confusion of the obtained text data candidate set after error correction through a language model, comparing and sequencing calculation results, and obtaining error-corrected text data.
3. The method as claimed in claim 2, wherein the step of preprocessing the text data after error correction to obtain a matrix vector of chinese question comprises:
removing target characters in the error-corrected text data to obtain preprocessed text data, wherein the target characters comprise preset characters and preset symbols;
performing word segmentation on the obtained preprocessed text data to obtain word segmentation processing results;
and vectorizing the word segmentation processing result by using a word vector tool to obtain a Chinese question matrix vector.
4. The method for classifying Chinese question sentences based on text error correction and neural networks as claimed in claim 3, wherein said step of inputting the matrix vectors of Chinese question sentences into a bidirectional gated cyclic unit layer to obtain intermediate semantic matrix vectors comprises:
inputting the matrix vector of the Chinese question sentence into a bidirectional gating circulation unit layer, wherein the bidirectional gating circulation unit layer comprises a plurality of gating circulation units, and a single word vector is input into each gating circulation unit layer according to the sequence of words in the Chinese question sentence;
and splicing the output of each gating circulation unit to generate an intermediate semantic matrix vector.
5. The method as claimed in claim 4, wherein the step of inputting the attention moment matrix vector to the convolutional neural network layer to obtain a global feature matrix vector comprises:
inputting the attention moment array vector into a convolutional neural network layer to enable the convolutional neural network layer to calculate convolutional kernels with different lengths, and obtaining a convolution result;
performing maximum pooling processing on the convolution result, and extracting local features;
and splicing the extracted local features to obtain a global feature matrix vector.
6. The method for classifying Chinese question sentences based on text error correction and neural network as claimed in claim 1, wherein said step of inputting said global feature matrix vector to a fully connected layer to obtain probability distribution of each category comprises:
inputting the global feature matrix vector to a full connection layer to obtain an output vector;
and converting the output vector into probability distribution of corresponding categories according to the normalized exponential function.
7. The method for classifying Chinese question sentences based on text error correction and neural network as claimed in claim 1, wherein the step of obtaining attention matrix vectors according to the attention weights corresponding to the intermediate semantic matrix vectors comprises:
calculating an attention value by adopting a normalization exponential function according to the output of the gating circulation unit at the current moment and the attention weight parameter;
calculating the attention vector at the current moment according to the current moment and the total length of the time sequence;
and splicing the attention vectors at different moments to obtain an attention moment array vector.
CN201910801515.4A 2019-08-28 2019-08-28 Chinese question classification method based on text error correction and neural network Active CN110516070B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201910801515.4A CN110516070B (en) 2019-08-28 2019-08-28 Chinese question classification method based on text error correction and neural network

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201910801515.4A CN110516070B (en) 2019-08-28 2019-08-28 Chinese question classification method based on text error correction and neural network

Publications (2)

Publication Number Publication Date
CN110516070A CN110516070A (en) 2019-11-29
CN110516070B true CN110516070B (en) 2022-09-30

Family

ID=68627740

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201910801515.4A Active CN110516070B (en) 2019-08-28 2019-08-28 Chinese question classification method based on text error correction and neural network

Country Status (1)

Country Link
CN (1) CN110516070B (en)

Families Citing this family (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111079416B (en) * 2019-12-03 2024-02-02 河海大学常州校区 Chinese text correction method based on shared control gate structure
CN111079410B (en) * 2019-12-23 2023-12-22 五八有限公司 Text recognition method, device, electronic equipment and storage medium
CN111159360B (en) * 2019-12-31 2022-12-02 合肥讯飞数码科技有限公司 Method and device for obtaining query topic classification model and query topic classification
CN111259658B (en) * 2020-02-05 2022-08-19 中国科学院计算技术研究所 General text classification method and system based on category dense vector representation
CN113468414A (en) * 2021-06-07 2021-10-01 广州华多网络科技有限公司 Commodity searching method and device, computer equipment and storage medium
CN113255332B (en) * 2021-07-15 2021-12-24 北京百度网讯科技有限公司 Training and text error correction method and device for text error correction model

Family Cites Families (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CA3055379C (en) * 2017-03-10 2023-02-21 Eduworks Corporation Automated tool for question generation
CN107797985B (en) * 2017-09-27 2022-02-25 百度在线网络技术(北京)有限公司 Method and device for establishing synonymous identification model and identifying synonymous text
CN108595602A (en) * 2018-04-20 2018-09-28 昆明理工大学 The question sentence file classification method combined with depth model based on shallow Model

Also Published As

Publication number Publication date
CN110516070A (en) 2019-11-29

Similar Documents

Publication Publication Date Title
CN110516070B (en) Chinese question classification method based on text error correction and neural network
CN108984526B (en) Document theme vector extraction method based on deep learning
CN110298037B (en) Convolutional neural network matching text recognition method based on enhanced attention mechanism
CN110245229B (en) Deep learning theme emotion classification method based on data enhancement
CN109376242B (en) Text classification method based on cyclic neural network variant and convolutional neural network
CN107943784B (en) Relationship extraction method based on generation of countermeasure network
CN109918491B (en) Intelligent customer service question matching method based on knowledge base self-learning
WO2023024412A1 (en) Visual question answering method and apparatus based on deep learning model, and medium and device
CN112100346B (en) Visual question-answering method based on fusion of fine-grained image features and external knowledge
CN111914097A (en) Entity extraction method and device based on attention mechanism and multi-level feature fusion
CN109977199B (en) Reading understanding method based on attention pooling mechanism
CN110232122A (en) A kind of Chinese Question Classification method based on text error correction and neural network
CN111966812B (en) Automatic question answering method based on dynamic word vector and storage medium
CN112052684A (en) Named entity identification method, device, equipment and storage medium for power metering
CN110019822B (en) Few-sample relation classification method and system
CN113255366B (en) Aspect-level text emotion analysis method based on heterogeneous graph neural network
CN111581364B (en) Chinese intelligent question-answer short text similarity calculation method oriented to medical field
CN113486645A (en) Text similarity detection method based on deep learning
CN111984791A (en) Long text classification method based on attention mechanism
CN113094502A (en) Multi-granularity takeaway user comment sentiment analysis method
CN110852071A (en) Knowledge point detection method, device, equipment and readable storage medium
CN108875024B (en) Text classification method and system, readable storage medium and electronic equipment
CN113806543A (en) Residual jump connection-based text classification method for gated cyclic unit
CN111723572B (en) Chinese short text correlation measurement method based on CNN convolutional layer and BilSTM
CN110334204B (en) Exercise similarity calculation recommendation method based on user records

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant