CN109241377B - Text document representation method and device based on deep learning topic information enhancement - Google Patents

Text document representation method and device based on deep learning topic information enhancement Download PDF

Info

Publication number
CN109241377B
CN109241377B CN201810999545.6A CN201810999545A CN109241377B CN 109241377 B CN109241377 B CN 109241377B CN 201810999545 A CN201810999545 A CN 201810999545A CN 109241377 B CN109241377 B CN 109241377B
Authority
CN
China
Prior art keywords
topic
document
information
matrix
vector
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN201810999545.6A
Other languages
Chinese (zh)
Other versions
CN109241377A (en
Inventor
张文跃
王素格
李德玉
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Shanxi University
Original Assignee
Shanxi University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Shanxi University filed Critical Shanxi University
Priority to CN201810999545.6A priority Critical patent/CN109241377B/en
Publication of CN109241377A publication Critical patent/CN109241377A/en
Application granted granted Critical
Publication of CN109241377B publication Critical patent/CN109241377B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Abstract

The invention discloses a text document representation method and device based on deep learning topic information enhancement. The method comprises the following steps: and S1, performing data preprocessing operation on the corpus document in the text form. And S2, designing a text sequence layer, and embedding the context information of each word in the word sequence in the expression vector of each word in the document. And S3, transitioning the sequence elements into the topic information of higher level through the attention layer. S4, in the topic layer, a representation of the current document D in all topic directions is generated. And S5, limiting the similarity degree between all topic information. And S6, fusing the topic representation vectors into a semantic representation vector Rep of the document D at a representation layer. S7, updating parameters of Rep through a classifier and an objective function, the method can effectively embed text sequence context semantic information and latent topic information into document representation vectors, and the topic information enhanced representation vectors can significantly improve the performance of a text mining model using the text sequence context semantic information and the latent topic information.

Description

Text document representation method and device based on deep learning topic information enhancement
Technical Field
The invention relates to the field of computer text representation learning, in particular to a text document representation method based on deep learning enhanced topic information enhancement and a text document representation device based on deep learning enhanced topic information enhancement.
Background
Document-level, holistic grasp on text is an important requirement for many text processing tasks. Currently, this problem is generally solved by textual representation learning. The text document level representation learning task mainly aims at constructing a method for converting a text document into a representation vector which can be directly operated by a computer according to the intrinsic semantic information of the text document. Specifically, a document in text form is represented as a fixed-length real-type vector that implies its semantics. Nowadays, document representation learning has become a fundamental and widespread application in the fields of natural language processing, text mining, information extraction, and the like.
The document representation learning methods that are currently most widely used are roughly of three categories, each of which has various disadvantages: (1) based on the "bag of words" (BoW) model, also known as the "vector space model". The expression vectors generated by the model are sparse and non-real, and the vectors are often not good in later application; (2) semantic analysis-based methods, such as a probabilistic latent semantic analysis model and an LDA document theme generation model, ignore context information of word sequences in texts, and restrict semantic carrying capacity of expression vectors; (3) the long-short term memory model (LSTM) based on the recurrent neural network is widely applied to the distributed representation vector generation of text documents. However, ordinary LSTM may not be sufficient to obtain global topic information of the corpus.
The disadvantages of the above approach show the difficulties currently faced by the document representation learning task: when the model is based on topic information of corpus globality, context information in documents is often lost (for example, the word "apple" cannot be determined without the context information to refer to fruit or science and technology companies), while when the model is focused on the local information, the globality topic information is ignored (correlation among documents), and in addition, no restriction mechanism exists among the topic information, so that the topic information tends to be similar, and the performance of the model is reduced (for example, a topic group with redundancy situations such as "economy", "entertainment", "chariot", "warship" is separated). All of these deficiencies can leave the representation vectors of the document lacking some semantic information, which can limit the effectiveness of these representation vectors in other applications at a later time.
Disclosure of Invention
The present invention is directed to solving, at least to some extent, one of the technical problems in the related art.
Therefore, an object of the present invention is to provide a text document representation method for enhancing topic information enhancement based on deep learning, which can enable a text document to generate a dense and real representation vector containing word order context information and topic information.
Another object of the present invention is to provide a text document presentation apparatus enhanced based on deep learning enhanced topic information.
In order to achieve the above object, an embodiment of the invention provides a text document representation method based on deep learning enhanced topic information enhancement, which includes the following steps:
s1, for a document D consisting of n words in a corpus containing K topics { w ═ w1,w2,...,wnCarrying out data preprocessing operations of cleaning, extracting, converting and sorting to obtain a word vector matrix D of the document (x)1,x2,...,xn};
S2, constructing a text sequence layer by using the sequence relation among the words, designing a sequence form long-term and short-term memory model, and acquiring a potential semantic matrix Hs (h) of the document1,h2,...,hnIn which hi=f1(xi,hi-1),h0=f1(x0),f1Operating for a neural network node;
s3, using the latent semantic matrix Hs ═ { h ═ h1,h2,...,hnGenerate a corresponding attention intensity matrix a ═ a1,a2,...,anAnd converting the A matrix and then normalizing the A matrix according to the rows to obtain an attention weight matrix A, wherein ai=f2(hi),f2Is a transformation function;
s4, fusing the latent semantic matrix Hs and the attention weight matrix A, and obtaining a mapping matrix expression VTs of all topics of the document, wherein VTs is f3(Hs, A), wherein f3Is a transformation function;
s5, using the label information of the cross-document to restrain the similarity degree of VTs represented by the mapping matrix of the topic, and obtaining a mapping matrix representation VTk after topic information enhancement;
s6, forVTk are fused to obtain a semantic representation vector Rep of the document D, wherein Rep is f4(VTk) wherein f4Is a fusion function;
s7, classifying the Rep by a topic classifier, obtaining an error index according to classification accuracy and topic similarity indexes, and updating the model parameters in the steps S1-S6 by using an objective function gradient descent method.
According to the text document representation method based on deep learning enhanced topic information enhancement provided by the embodiment of the invention, firstly, words in a text form are converted into word vectors by using a word embedding technology, so that the form of a document is changed into a real number matrix, and then a text sequence layer is established according to the characteristic of text context semantic information sequentiality. After the real matrix of the document passes through the sequence layer, the real matrix becomes a potential semantic matrix with context semantic information. And then, calculating an attention weight matrix corresponding to the latent semantic matrix by using the latent semantic matrix, and realizing the enhancement of topic information with higher granularity by fusing the latent semantic matrix and the attention weight matrix. Then, topic similarity constraint mechanism is used to make the topics be distinguished from each other as much as possible, so as to obtain all topic representations of the document. And finally, fusing all topic representations to serve as the enhanced representation vector of the topic information of the document, so that the text document is a dense and real representation vector containing word sequence context information and topic information, and the topic redundancy is reduced.
In order to achieve the above object, another embodiment of the present invention provides a text document representation apparatus enhanced based on deep learning enhanced topic information, including a text sequence layer, an attention layer, a topic layer and a representation layer, wherein the text sequence layer is used for D ═ w { of n words in a corpus containing K topics1,w2,...,wnCarrying out data preprocessing operations of cleaning, extracting, converting and sorting to obtain a word vector matrix D of the document (x)1,x2,...,xnAnd the word vector matrix D of the document is set as { x }1,x2,...,xnAcquiring a potential semantic matrix Hs (h) of the document through a sequence form long-term and short-term memory model1,h2,...,hnIn which hi=f1(xi,hi-1),h0=f1(x0),f1Operating for a neural network node; the attention layer is used for extracting and separating topic information in the text, realizing connection of two kinds of granularity information from a word level to a topic level and realizing a function of extracting unknown information from known information; from the latent semantic matrix Hs ═ h1,h2,...,hnGenerate a corresponding attention intensity matrix a ═ a1,a2,...,anAnd converting the A matrix and then normalizing the A matrix according to the rows to obtain an attention weight matrix A, wherein ai=f2(hi),f2The topic layer is used for obtaining mapping matrix expression VTs of all topics of the document, and the VTs is f3(Hs, A), wherein f3Is a transformation function; and the similarity degree of the mapping matrix representation VTs of the topic is restrained by using the label information of the cross-document, and the mapping matrix representation VTk after topic information enhancement is obtained; the representation layer is used for fusing the VTk to obtain a semantic representation vector Rep of the document D, wherein Rep is f4(VTk) wherein f4And fusing functions, classifying the Rep by a topic classifier, obtaining an error index according to classification accuracy and topic similarity indexes, and updating model parameters by using a target function gradient descent method.
According to the text document representation device based on deep learning enhanced topic information enhancement provided by the embodiment of the invention, firstly, words in a text form are converted into word vectors by using a word embedding technology, so that the form of a document is changed into a real number matrix, and then a text sequence layer is established according to the characteristic of text context semantic information sequentiality. After the real matrix of the document passes through the sequence layer, the real matrix becomes a potential semantic matrix with context semantic information. And then calculating an attention weight matrix corresponding to the latent semantic matrix at the attention layer by using the latent semantic matrix, and realizing the enhancement of topic information with higher granularity by fusing the latent semantic matrix and the attention weight matrix. Then, topic similarity constraint mechanism is used at topic layer to make the topics be distinguished from each other as much as possible, so as to obtain all topic representations of the document. And finally, fusing all topic representations to serve as the enhanced representation vector of the topic information of the document, so that the text document is a dense and real representation vector containing word sequence context information and topic information, and the topic redundancy is reduced.
Compared with the prior art, the invention has the following beneficial effects:
1. a sequence LSTM model is adopted to model the word sequence of the text, so that the model can better fuse the context information of the text;
2. the extraction-type attention mechanism of the entirely new structure supports a "sequence-to-tree" structure process, which is used to extract topic information from text sequence information. In addition, the method not only can embed the word-topic associated information of the text into a representation vector, but also can explicitly return the support degree of words in the document to different topics, and can be used as a visual result for displaying and testing;
3. the introduction of the similarity constraint mechanism of the topic layer improves the 'long tail effect' of the original topic model, namely, the model is degraded due to the fact that some topics are too similar. Meanwhile, the problem of the convergence faced by the general attention mechanism is also solved. The tendency is caused by too few variables in the process of calculating attention, so that the attention weight distribution of all topics tends to be the same, and a similarity constraint mechanism adds the variables to the calculation process;
4. the novel document expression vector generation method is formed by combining a plurality of special sub-models, and in the whole view, the models can not only encode the context semantic information in the local document, but also enhance the global latent topic semantic information at the corpus level so as to be embedded into the final document expression vector;
5. the innovation point of the method is that various innovation submodels are designed according to different semantic information and are compounded into a deep learning model for document representation learning. The most important innovation is the attention mechanism of the sequence-to-tree structure and the design of the topic information similarity constraint mechanism. Experiments on different data sets show that the document expression vector generated by the method is superior to other classical comparison models in three main text mining tasks of text classification, topic identification and text clustering, and the method can practically improve the quality of the text expression vector.
Drawings
FIG. 1 is an overall hierarchy framework diagram of the present invention.
FIG. 2 is a view of the attention layer structure described in steps S3-S4.
Fig. 3 is a schematic diagram of the topic similarity constraint mechanism in step S5.
FIG. 4A is a comparison of results of various algorithm-generated document representation vectors in a classification experiment.
Fig. 4B is a visualization display of the relevance of topic difference and document classification accuracy.
FIG. 5 is a visual display of the effect of the invention on the topic identification task.
FIG. 6 is a comparison of the results of the present invention with classical algorithms in a text clustering task.
FIG. 7 is a flowchart of a text document representation method based on deep learning topic information enhancement according to the present invention.
Detailed Description
In this embodiment, the experiment of the text document representation method based on deep learning topic information enhancement of the present invention is completed on a computer of Shanxi university and a cluster computer of information technology institute, the cluster is composed of 5 high performance computers to form a computing and managing node, and the network connection adopts gigabit Ethernet and infiniband 2.5G network. Each node is provided with an eight-core CPU and a 128GB memory, the CPU is intel xeon E3-1230V53.4GMhz main frequency and is provided with two NVIDIA GTX1080 high-performance graphics cards, and large-scale matrix operation and deep learning model training can be carried out.
As can be seen from fig. 1 to 7, the present invention divides different semantic information into several sub-models, which are connected layer by layer and finally fused. The learning process mainly comprises the following steps:
s1, for a document D consisting of n words in a corpus containing K topics { w ═ w1,w2,...,wnCleaning, extracting and rotatingPreprocessing the data to obtain the word vector matrix D ═ x of the document1,x2,...,xnThe method comprises the following specific steps:
and S11, extracting and cleaning all text data, wherein if the text data is English data, marking, word stem and the like are needed, and if the text data is Chinese data, Chinese word segmentation is needed. And removing stop words in the data and deleting the document with too few words (less than 6 words).
S12, converting all words in the corpus into Word vectors by using the Word2Vec Word vector model after being pre-trained by the big corpus. Where the obscure words (not present in the word vector model) are discarded.
S13, acquiring labels of the training corpus, wherein K are respectively corresponding to K topics, and each topic is corresponding to a unique one-hot form vector for supervising the learning process. These tag vectors and the document data after the preprocessing thereof are associated with each other as experimental data.
S2, for potential semantic extraction of context, the invention constructs a text sequence layer by using sequence relation among words, designs a sequence form long-short term memory model (seq-LSTM), and embeds context information of each word in the word sequence in the expression vector of each word in the document. The method comprises the following specific steps:
s21, calculating states of all gate elements of the LSTM, wherein the LSTM gate elements play a control role in calculation, are flexibly adjusted according to input information and are mainly divided into an input gate, an output gate and a forgetting gate, and the input and the output of deep learning node information and the adjustment of historical information are respectively controlled, and the specific calculation mode is as follows:
Figure BDA0001782637400000061
i, F, O and G are input gates, output gates, forgetting gates and node information states respectively, sigma represents a sigmoid activation function, tanh is a hyperbolic tangent function, Wseq and Bseq are weight matrixes and offset vectors of the deep learning neural network respectively, and seq represents that parameters belong to a text sequence layer. All the gate states are calculated by historical information and current word vector input according to the formula;
s22, calculating the LSTM hidden state. The hidden state is a module in the long-short term memory model for storing history or other information, and is calculated as follows:
Ct=It·Gt+Ft·Ct-1
wherein C represents a node hidden state corresponding to a certain word, and the hidden state is influenced by node information and a historical hidden state which are respectively regulated by an input gate and a forgetting gate, and the regulation is realized by element multiplication among vectors. In a word, the hidden state of the current word is balanced and adjusted between the current input state and the historical state according to semantic information;
and S23, calculating the LSTM node state. After obtaining the hidden state corresponding to the current word of the document, activating the hidden state to obtain the latent context semantic state corresponding to the word:
ht=Ot·tanh(Ct)
as shown in the formula, the activation function selects the hyperbolic tangent function, and the activation value is used as the node state for subsequent calculation after being adjusted by the output gate.
And S24, recording the text sequence layer result. Document D ═ { x1,x2,...,xnH, generating a corresponding semantic state matrix Hs (h) through a text sequence layer1,h2,...,hnC and a hidden state matrix Cs ═ C1,C2,...,CnThe two matrixes contain context semantic information in the document D, for example, vectors of the words "cry" in "happy to cry" and "cry" in "sad to cry" are the same, but the expression vectors (node states h) of the two words are different due to the difference of the two words after the sequence layer.
S3, in order to enhance the topic information in the context semantic information of the document, the sequence elements are necessarily transited into the topic information of higher level, for which the invention proposes a new extraction type attention mechanism and constructs on the text sequence layer, as shown in FIG. 2. While the past attention mechanism often concatenated two sequence structures, the present invention requires concatenating sequences and tree nodes, where each sequence element represents a position in the sequence of document words and each tree node represents a topic. And it is generally noted that both structures in the mechanism are known information, whereas the extraction mechanism of the present invention extracts potential information (i.e., topics) from the known information. The method comprises the following specific steps:
and S31, acquiring attention intensity. The attention intensity is calculated according to the following formula according to the document context semantic information:
Figure BDA0001782637400000071
wherein WattAnd battWeight matrix and offset vector parameters, a, of the attention layer, respectivelytIs a K-dimensional vector, and the value of each dimension of the K-dimensional vector represents the attention intensity of the t-th word of the document to the corresponding topic.
And S32, calculating an attention weight matrix. The attention intensity matrix a obtained after the step S31 is an n × K matrix, which is first transformed into K × n, so that the meaning of the matrix becomes the attention (expression) intensity of each line indicating the topic of the current document text sequence, for example: the word "apple" at a certain position in a document expresses topic 1, topic 2 … …, and the like to a certain extent (unlike the conventional attention mechanism, the specific contents of the topics are not required to be specified, and are unknown).
Figure BDA0001782637400000072
This intensity distribution is then normalized to the form of a probability distribution by the following softmax algorithm:
Figure BDA0001782637400000073
finally, recording the normalized attention weight matrix A:
Figure BDA0001782637400000074
s4, in the topic layer, the attention weight matrix A from the attention layer is fused with the document context semantic information Hs generated by the text sequence layer. The semantic information Hs is merged according to the corresponding different attention weights, and because the expression of the weight reaction expression vector to the topic is strong and weak, the potential topic information in the original semantic information is enhanced or emphasized. Finally, a representation of the current document D in all topic directions is generated, which can also be regarded as a mapping of its semantic information in all topic spaces (which can be understood as what an article about apple company looks like in different topic perspectives such as "science", "economy", "politics"). As shown in fig. 1 and 2, the model has K topics representing nodes VTs corresponding to all topics in the corpus, and VTCs are hidden states of VTs generated by using LSTM type deep learning nodes, and they are calculated as follows:
Figure BDA0001782637400000081
the VTs and the VTCs are provided with K rows, and each row vector corresponds to a topic related information representation vector and also corresponds to an LSTM type node. The topic representation of the document D is obtained by weighting and summing all the context semantic information of the document D according to the expression intensity of the topic representation.
According to the characteristics of the global topic information and the local context semantic information of the text, a plurality of substructures are designed, and are laminated and compounded, and then the overall structure is used for learning document semantic representation. Due to the design, different types of semantic information can be processed with corresponding modules in a targeted mode, and due to the fact that different types of information are different greatly, integration of different modules cannot be achieved simply through stacking, and therefore the design is provided with an extraction type attention mechanism to be responsible for bridging between the semantic modules and the topic modules.
S5, constraining the degree of similarity between all VTs. As described above, there may be a tendency of convergence between the previous model-generated topic representations, for example, there should be a "military" topic in the corpus, but the model is decomposed into topics such as "weapons" and "military", and other topics should be merged, which often appears in the corpus with a large difference in the number of topic documents. The problem is reflected in the model of the invention, namely, VT expression vectors are mathematically too similar, so that K topic information can be obviously lost, and the performance of the model is degraded. Therefore, at the topic level, the invention designs a unique topic information similarity constraint mechanism as shown in fig. 3. Wherein, L is a topic label vector with the length of K and the form of 'one-hot' (the value of a certain position is close to 1, and the rest positions are close to 0), the basic principle of the constraint mechanism is that a contrast vector v generated by the topic representation vector is gradually similar to the label L through the training process, and the difference degree between the topic information representation vectors can be gradually enlarged due to the high orthogonality between the L. The concrete implementation steps of the similarity constraint mechanism are as follows:
and S51, converting topic expression vectors. The dimensions of the topic representation vectors VT and VTC are not necessarily equal to K and are computationally infeasible to compare with the topic label L, so their lengths are first transformed by the following algorithm:
Figure BDA0001782637400000082
ws and Bs are weight matrix parameters and bias matrix parameters in a topic information similarity constraint mechanism, sigma is a sigmoid activation function, the length of a contrast vector vk is K, and each document has K contrast vectors and each vector corresponds to one topic.
And S52, similarity measurement. The invention adopts the cross entropy as the similarity measurement between the contrast vector and the topic label vector, and the calculation mode is as follows:
Figure BDA0001782637400000091
when s iskSmaller values represent contrast vectors vkAnd topic tag vector LkThe more similar, the generation of v is now demonstratedkTopic information vector VT ofkAnd VTCkThe greater the difference with other topic vectors.
And S53, calculating topic similarity scores. After the similarity scores of all topics are obtained, averaging the similarity scores to obtain topic information, wherein the similarity comprehensive score S is as follows:
Figure BDA0001782637400000092
the smaller the S value is, the smaller the similarity between topic information representation vectors is, the smaller the topic information redundancy is, and the more comprehensive the topic information in the document representation vectors generated by the invention is. The present invention minimizes the S value by objective function error feedback and parameter update during the training phase.
And S6, fusing the topic representation vectors into a semantic representation vector Rep of the document D at a representation layer. In step S5, K topic information expression vectors are obtained, the topic expression vectors are used as leaf nodes of a tree and the final document expression vector Rep is used as a parent node in a presentation layer through a tree LSTM model, and semantic information is gathered from child nodes to the parent node through LSTM type operation, which specifically comprises the following steps:
s61, tree LSTM gate element state calculation. The input gate, output gate and node state of tree LSTM father node are calculated first, the algorithm is slightly different from the previous sequence part:
Figure BDA0001782637400000093
wherein Wtr、BtrAnd indicating that the weight matrix and the bias matrix are positioned at the representation layer of the tree type. As can be seen from the equation, the data in the K representative vectors are integrated to generate a single gate element, no longer having a differenceTopic differentiation, as all enhanced topic information is already contained in the final state vectors I, O and G.
And S62, calculating the special forgetting door state. Unlike the rest of the gate elements, the forgetting gate in the tree structure of the present invention plays a role of controlling the information flow from the child node to the parent node, so that each child node has one forgetting gate, and the calculation of the forgetting gates between nodes is independent from each other (due to the independence between the subject information). For example, a specific algorithm for forgetting the state of the gate by the kth topic child node is as follows:
Figure BDA0001782637400000101
the expression shows that the forgetting gate state of the corresponding child node of each topic is calculated by the semantic information contained in the topic expression vector.
And S63, hidden state calculation. The hidden state of the LSTM node in the sequence structure stores history information and the hidden state of the parent node in the tree structure stores information from child nodes that are controlled by their respective forgetting gates to reach the parent node, as mentioned in step F2. When the parent node calculates its hidden state, the adjusted child node information is combined as follows:
Figure BDA0001782637400000102
and S64, generating a document representation vector. In the step, firstly, the hidden state of the father node is processed by an activation function and an output gate to obtain a node state vector, and finally, a representation vector Rep of the current document is obtained by one-layer dimension adjustment. The specific calculation method is as follows:
h=O·tanh(C)
Rep=σ(Wrh+br)
wherein, WrAnd BrAre parameters of a deep learning neural network. Due to document representationThe required length of the vector may not be consistent with the dimension of the deep learning hidden layer, so the invention additionally adds a vector length adjusting operation.
S7, classifier layer and objective function. In order to train the model of the invention, after semantic expression vectors of the document are obtained, the vectors are classified through a topic classifier, the classification accuracy is recorded, a topic similarity index is added to obtain a system error index of the document D of the current model, and then the model parameters of the invention are updated by an error feedback algorithm of a deep learning model and a target function gradient descent method. The objective function of the present invention is as follows:
Figure BDA0001782637400000103
and b, adjusting the balance of classification precision and topic difference degree by the lambda parameter, wherein g is a topic category mark of the document D, and p is a classification result made by the classifier according to the document Rep.
The text expression vector generated by a good expression learning method can make a natural language processing task applying the vector perform better because the text expression vector contains more and more accurate semantic information, so that the invention tests the generated document expression vector by using three tasks of text classification, topic detection and text clustering which are most widely applied.
Fig. 4A and 4B are experimental performances of the document expression vectors generated by the present invention in topic classification, which are a classification precision experiment and a topic information similarity validity experiment, respectively. To verify the classification performance of the representative vectors, the experiment used three classes of text corpora, with 90% of the documents in the corpora used for training and the remainder used for testing. The choice word vector dimension, the deep learning hidden layer dimension, and the representation vector dimension are 50, 100, and 50, respectively. The parameter λ of the objective function is 0.2, the initial value of the model learning rate is 0.1, and the learning method is adagard. Referring to fig. 4A, almost on all corpora, the accuracy of the invention (TE-LSTM) is better than that of other classical contrast algorithms, and the result with the topic information similarity constraint mechanism (with SC) is better than that without the mechanism (with SC), which shows that the representation learning method proposed by the invention can improve the semantic information amount in the representation vector, and the topic information similarity constraint mechanism obviously plays a positive role. In fig. 4B, the abscissa indicates the degree of difference between topic information, the larger the numerical value is, the lower the topic information similarity is, and the ordinate of the graph indicates the classification accuracy of the document in the degree of difference interval. As can be seen from the curve of fig. 4B, as the difference between the topic information is larger, the classification accuracy of the expression vector also shows a trend of increasing gradually, which also illustrates the effectiveness of the topic information similarity constraint mechanism of the present invention, which reduces the topic information redundancy of the model and improves the information expression capability of the vector.
FIG. 5 shows the presentation of document representation vectors generated by the present invention in a topic identification task. The left-most side in the table is the model name, which is lda2vec, the invention (no topic information constraint mechanism) and the invention (with constraint mechanism) model. The second column is the topic tag in the corpus, listing 4 of the 20 topics. The third column is keywords of topics detected from the corpus, the keywords are words with a degree of criticality ranking 5 in each topic obtained by model calculation, and the degree of criticality of a word in the invention is the attention weight value of the word to the topic. The values in the last column are topic relevance calculated after the online platform Palmetto receives 5 keywords, and the higher score indicates that the semantics of the keywords are close to that they are more likely to originate from the same topic. The analysis chart shows that the method obtains obviously better experiment results from both qualitative and quantitative angles, and the similar classification experiment proves that the quality of the expression vector is improved by all designs of the method, because the model performance of the topic information similarity constraint mechanism is more excellent.
FIG. 6 shows the representation of the document semantic information representation vector generated by the present invention in the text clustering task. Representation learning is a task of converting text form data into a representation vector which can be directly calculated, and semantic information of text can be intuitively reflected through calculation, for example, the closer words in a word vector are, the smaller the vector distance between the words is, the more closely the words are. Similarly, the degree of correlation between the documents can also be determined by calculating the distance between the document representation vectors. The better the quality, the higher the correlation between the degree of relevance of the representation vector document and the vector distance. Accordingly, a text clustering task is set to detect the performance of the vectors generated by the invention. The more documents belonging to the same topic are clustered and judged in the same cluster, the better the vector performance is proved, and the representation learning of the invention is more excellent. The numerical values in fig. 6 are calculated in the following manner: calculating the document content of the most topics in the clusters one by one, recording the document content if the topic does not have a corresponding cluster, selecting the topic with the second highest content until meeting the topic which is not arranged, and taking the average value of the document content of all the topics as the text cluster score of the model after all the clusters have the corresponding topics. Referring to fig. 6, it is known that the expression vector of the present invention has the best clustering effect, and the model using the topic information similarity constraint mechanism obtains the highest score, which proves that the present invention can generate the document semantic expression vector with better quality.
In summary, for a document D consisting of n words in a corpus containing K topics, { w ═ w1,w2,...,wnThe invention adopts the following technical scheme:
obtaining word vector matrix D ═ { x) of document through pre-training1,x2,...,xnIn the context sequence, the potential semantic h corresponding to each wordi=f1(xi,hi-1),h0=f1(x0) Where f is the conversion function. Thus, even if the same word has different latent semantics in different context (i.e. the same word has different representation vectors in different positions in the text), the difference is just the proof of the implication of the context semantic information. Further, f in the formula1It may be a neural network node operation.
In topic information acquisition. From the underlying semantic matrix H ═ H of the document1,h2,...,hnGenerate a corresponding attention intensity matrix a ═ a1,a2,...,an} of whichIn (a)i=f2(hi) Is that each dimension of the K-dimension vector represents the attention intensity (or expression intensity) of the ith word in the sequence to a certain topic, f2Is a conversion function. And finally, the A matrix is rotated and then is normalized according to rows to obtain an attention weight matrix A.
In the aspect of topic information enhancement, the context semantic information and the attention weight of the document are combined to generate a mapping matrix (VT) of all topics of the document, wherein the VT is f3(H, A) wherein f3Each line of VT corresponds to a topic, and represents the information of the topic contained in the document D. After this portion, the topic information of the document has been individually enhanced.
In the aspect of topic information control, the document topic information obtained in the last stage is limited by using label information of the whole corpus of the cross-document. Each topic has its fixed corresponding label vector L, e.g., LiThis vector is used to pair VTiAnd limiting, namely controlling the neural network classifier by using similar supervision information, wherein L is the supervision information, label vectors corresponding to each topic are highly orthogonal, and the topic information limited by the labels is highly different from each other naturally.
And fusing the semantic information after topic enhancement into a document expression vector. Topics are not related to each other and are all attributed to a representation vector, and a typical tree structure is formed. Different from the common way of combining according to the weight, the lack of the weight requires that all topic vectors are fused in a more comprehensive way, and the fusion way is set as f4If the expression vector of the document D is Rep, then Rep ═ f4(VT). During training, a classifier is arranged on the Rep, the classifier is trained by using the class vector of the document, and the model is updated by using an error return and gradient descent mode.
Therefore, the text document is a dense and real expression vector containing word sequence context information and topic information, and topic redundancy is reduced.
In the description of the present invention, it is to be understood that the terms "central," "longitudinal," "lateral," "length," "width," "thickness," "upper," "lower," "front," "rear," "left," "right," "vertical," "horizontal," "top," "bottom," "inner," "outer," "clockwise," "counterclockwise," "axial," "radial," "circumferential," and the like are used in the orientations and positional relationships indicated in the drawings for convenience in describing the invention and to simplify the description, and are not intended to indicate or imply that the referenced device or element must have a particular orientation, be constructed and operated in a particular orientation, and are not to be considered limiting of the invention.
For this reason, the terms "first", "second" are used for descriptive purposes only and are not to be construed as indicating or implying relative importance or implicitly indicating the number of technical features indicated. Thus, a feature defined as "first" or "second" may explicitly or implicitly include at least one such feature. In the description of the present invention, "a plurality" means at least two, e.g., two, etc., unless explicitly specified otherwise.
In the present invention, unless otherwise expressly stated or limited, the terms "mounted," "connected," "secured," and the like are to be construed broadly and can, for example, be fixedly connected, detachably connected, or integrally formed; can be mechanically or electrically connected; they may be directly connected or indirectly connected through intervening media, or they may be connected internally or in interactive relationship with each other unless otherwise specifically limited. The specific meanings of the above terms in the present invention can be understood by those skilled in the art according to specific situations.
In the present invention, unless otherwise expressly stated or limited, the first feature "on" or "under" the second feature may be directly contacting the first and second features or indirectly contacting the first and second features through an intermediate. Also, a first feature "on," "over," and "above" a second feature may be directly on or obliquely above the second feature, or simply mean that the first feature is at a lesser level than the second feature.
In the description herein, references to the description of the term "one embodiment," "some embodiments," "an example," "a specific example" or "some examples" or the like are intended to mean that a particular feature, structure, material, or characteristic described in connection with the embodiment or example is included in at least one embodiment or example of the invention. In this specification, the schematic representations of the terms used above are not necessarily intended to refer to the same embodiment or example. Furthermore, the particular features, structures, materials, or characteristics described may be combined in any suitable manner in any one or more embodiments or examples. Furthermore, various embodiments or examples and features of different embodiments or examples described in this specification can be combined and combined by one skilled in the art without contradiction.
The accompanying drawings and the detailed description are included to provide a further understanding of the invention. The method of the present invention is not limited to the examples described in the specific embodiments, and other embodiments derived from the method and idea of the present invention by those skilled in the art also belong to the technical innovation scope of the present invention. This summary should not be construed to limit the present invention.

Claims (8)

1. A text document representation method based on deep learning topic information enhancement is characterized by comprising the following steps:
s1, for a document D consisting of n words in a corpus containing K topics { w ═ w1,w2,...,wnCarrying out data preprocessing operations of cleaning, extracting, converting and sorting to obtain a word vector matrix D of the document (x)1,x2,...,xn};
S2, constructing a text sequence layer by using the sequence relation among the words, designing a sequence form long-term and short-term memory model, and acquiring a potential semantic matrix Hs (h) of the document1,h2,...,hnIn which hi=f1(xi,hi-1),h0=f1(x0),f1Operating for a neural network node;
s3, using the latent semantic matrix Hs ═ { h ═ h1,h2,...,hnGenerate a corresponding attention intensity matrix a ═ a1,a2,...,anAnd converting the A matrix and then normalizing the A matrix according to the rows to obtain an attention weight matrix A, wherein ai=f2(hi),f2Is a transformation function;
s4, fusing the latent semantic matrix Hs and the attention weight matrix A, and obtaining a mapping matrix expression VTs of all topics of the document, wherein VTs is f3(Hs, A), wherein f3Is a transformation function;
s5, using the label information of the cross-document to restrain the similarity degree of VTs represented by the mapping matrix of the topic, and obtaining a mapping matrix representation VTk after topic information enhancement;
s5 includes the steps of:
s51, topic representation vector transformation, the dimensionality of topic representation vectors VT and VTC are not necessarily equal to K, so their lengths are first transformed by the following algorithm:
Figure FDA0002947475920000011
ws and Bs are weight matrix parameters and bias matrix parameters in a topic information similarity constraint mechanism, sigma is a sigmoid activation function, the length of a contrast vector vk is K, and each document has K contrast vectors and each vector corresponds to a topic;
s52, measuring the similarity, wherein the cross entropy is used as the similarity measurement between the contrast vector and the topic label vector, and the calculation method is as follows:
Figure FDA0002947475920000012
when the sk value is smaller, the comparison vector vk and the topic label vector Lk are more similar, the comparison difference between the topic information vector VTk and the VTCk which generate vk and other topic vectors is proved to be larger, wherein L is the topic label vector with the length of K in the form of 'one-hot'; k training corpus topic labels are respectively corresponding to K topics, and each topic is corresponding to a unique one-hot form vector for supervising the learning process; the label vectors and the preprocessed document data are mutually corresponding to serve as experimental data;
s53, calculating topic similarity scores, averaging the similarity scores of all topics to obtain a comprehensive topic information similarity score S:
Figure FDA0002947475920000021
the smaller the S value is, the smaller the redundancy of topic information is, and the more comprehensive the topic information in the generated document expression vector is; minimizing the S value through target function error feedback and parameter updating in a training stage;
s6, fusing the VTk to obtain a semantic expression vector Rep of the document D, wherein Rep is f4(VTk) wherein f4Is a fusion function;
s7, classifying the Rep by a topic classifier, obtaining an error index according to classification accuracy and topic similarity indexes, and updating the model parameters in the steps S1-S6 by using an objective function gradient descent method.
2. The text document representation method enhanced based on deep learning topic information as claimed in claim 1, wherein S1 comprises the steps of:
s11, extracting and cleaning all text data, wherein if the text data is English data, marking and stemming are carried out; if the data is Chinese data, performing Chinese word segmentation processing; removing stop words in the text data, and deleting documents with the number of words less than six words;
s12, converting all words in the corpus into Word vectors by using the Word2Vec Word vector model after being pre-trained by the big corpus.
3. The text document representation method enhanced based on deep learning topic information as claimed in claim 1, wherein S2 comprises the steps of:
s21, designing a sequence form long-term and short-term memory model, namely an LSTM model, and calculating in the following way,
Figure FDA0002947475920000022
i, F, O and G are respectively an input gate, an output gate, a forgetting gate and a node information state, sigma represents a sigmoid activation function, tanh is a hyperbolic tangent function, Wseq is a weight matrix of the deep learning neural network, Bseq is a bias vector of the deep learning neural network, and seq represents that a parameter belongs to a text sequence layer;
s22, calculating the hidden state Ct corresponding to the current word of the document according to the LSTM model in the following way,
Ct=It·Gt+Ft·Ct-1
s23, according to LSTM model and the hidden state Ct corresponding to the current word of the document, activating the hidden state Ct to obtain the potential context semantic state corresponding to the word, the calculation mode is as follows,
ht=Ot·tanh(Ct)
s24, recording the text sequence layer result, document D ═ x1,x2,...,xnH, generating a corresponding semantic state matrix Hs (h) through a text sequence layer1,h2,...,hnC and a hidden state matrix Cs ═ C1,C2,...,CnAnd the two matrixes contain context semantic information in the document D.
4. The text document representation method enhanced based on deep learning topic information as claimed in claim 3, wherein S3 comprises the steps of:
s31, obtaining notes according to the context semantic information of the document DStrength of gravity atThe calculation method is as follows,
Figure FDA0002947475920000031
wherein, atIs a K-dimensional vector representing the attention intensity of the tth word of the document to the corresponding topic, WattAnd battRespectively are a weight matrix and a bias vector parameter of the attention layer;
s32, calculating an attention weight matrix; the attention intensity matrix a obtained in step S31 is ═ a1,a2,...,anIs an n x K matrix, which is first transformed into K x n, i.e.,
Figure FDA0002947475920000032
this intensity distribution is normalized to the form of a probability distribution by the following softmax algorithm,
Figure FDA0002947475920000033
the normalized attention weight matrix a is finally recorded as follows,
Figure FDA0002947475920000034
5. the text document representation method enhanced based on deep learning topic information as claimed in claim 4, wherein S4 comprises the steps of:
fusion is realized through a potential semantic matrix Hs and the attention weight matrix A, and mapping matrix representation of the current document D on all topics is obtained; VTs corresponds to all K topics in the corpus, and VTCs is a hidden state corresponding to VTs, and their calculation methods are as follows:
Figure FDA0002947475920000041
the VTs and the VTCs both have K lines, and each line vector corresponds to a related information representation vector of a topic.
6. The text document representation method enhanced based on deep learning topic information as claimed in claim 5, wherein S6 comprises the steps of:
s61, calculating the states of the input gate, the output gate and the node of the tree LSTM father node by the following method,
Figure FDA0002947475920000042
wherein Wtr、BtrThe weight matrix and the bias matrix are represented in a tree-shaped representation layer, and as can be seen from the formula, data in the K representation vectors are integrated to generate a single gate element, so that no distinction of different topics is needed any more, because all enhanced topic information is already contained in the final state vectors I, O and G;
s62, calculating the special forgetting gate state, different from other gate elements, each child node in the tree LSTM model structure has a forgetting gate, and the forgetting gate calculation among the nodes is also independent, wherein the forgetting gate plays a role in controlling the information flow condition from the child node to the father node, the forgetting gate state of the child node of the kth topic is calculated in the following way,
Figure FDA0002947475920000043
s63, calculating hidden state, storing information from child nodes in the hidden state of father node in the tree LSTM model structure, combining the adjusted information of child nodes when father node calculates the hidden state, calculating the hidden state by the following method,
Figure FDA0002947475920000044
s64, generating a document expression vector, firstly obtaining a node state vector from the hidden state of a father node through an activation function and an output gate, and finally obtaining an expression vector Rep of the current document through one-layer dimension adjustment, wherein the specific calculation method is as follows:
h=O·tanh(C)
Rep=σ(Wrh+br)
wherein, WrAnd BrAre parameters of a deep learning neural network.
7. The text document representation method enhanced based on deep learning topic information as claimed in claim 6, wherein S7 comprises the steps of:
setting a classifier and an objective function, recording a classification result of a semantic expression vector Rep of a document through a topic classifier, adding a topic similarity index to obtain a system error index of a current document D, then updating model parameters by using an objective function gradient descent method through a deep learning error feedback algorithm, wherein the objective function is shown as follows,
Figure FDA0002947475920000051
wherein, the lambda parameter adjusts and balances the classification precision and the topic difference degree, g is the topic category mark of the document D, and p is the classification result.
8. A text document presentation apparatus enhanced based on deep learning topic information, comprising:
a text sequence layer for matching a document D consisting of n words in a corpus containing K topics { w ═ w1,w2,...,wnCarrying out data preprocessing operations of cleaning, extracting, converting and sorting to obtain a word vector matrix D of the document (x)1,x2,...,xnAnd the word vector matrix D of the document is set as { x }1,x2,...,xnAcquiring a potential semantic matrix Hs (h) of the document through a sequence form long-term and short-term memory model1,h2,...,hnIn which hi=f1(xi,hi-1),h0=f1(x0),f1Operating for a neural network node;
the attention layer is used for extracting and separating topic information in the text, realizing connection of two kinds of granularity information from a word level to a topic level and realizing a function of extracting unknown information from known information; from the latent semantic matrix Hs ═ h1,h2,...,hnGenerate a corresponding attention intensity matrix a ═ a1,a2,...,anAnd converting the A matrix and then normalizing the A matrix according to the rows to obtain an attention weight matrix A, wherein ai=f2(hi),f2Is a conversion function, and realizes the fusion of the latent semantic matrix Hs and the attention weight matrix A,
a topic layer for obtaining mapping matrix expression VTs of all topics of the document, wherein VTs is f3(Hs, A), wherein f3Is a transformation function; and the similarity degree of the mapping matrix representation VTs of the topic is restrained by using the label information of the cross-document, and the mapping matrix representation VTk after topic information enhancement is obtained;
a presentation layer, configured to fuse the VTk to obtain a semantic representation vector Rep of the document D, where Rep ═ f4(VTk) wherein f4And fusing functions, classifying the Rep by a topic classifier, obtaining an error index according to classification accuracy and topic similarity indexes, and updating model parameters by using a target function gradient descent method.
CN201810999545.6A 2018-08-30 2018-08-30 Text document representation method and device based on deep learning topic information enhancement Active CN109241377B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201810999545.6A CN109241377B (en) 2018-08-30 2018-08-30 Text document representation method and device based on deep learning topic information enhancement

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201810999545.6A CN109241377B (en) 2018-08-30 2018-08-30 Text document representation method and device based on deep learning topic information enhancement

Publications (2)

Publication Number Publication Date
CN109241377A CN109241377A (en) 2019-01-18
CN109241377B true CN109241377B (en) 2021-04-23

Family

ID=65069456

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201810999545.6A Active CN109241377B (en) 2018-08-30 2018-08-30 Text document representation method and device based on deep learning topic information enhancement

Country Status (1)

Country Link
CN (1) CN109241377B (en)

Families Citing this family (12)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110135551B (en) * 2019-05-15 2020-07-21 西南交通大学 Robot chatting method based on word vector and recurrent neural network
CN110298038B (en) * 2019-06-14 2022-12-06 北京奇艺世纪科技有限公司 Text scoring method and device
CN110489563B (en) * 2019-07-22 2022-08-23 北京百度网讯科技有限公司 Method, device, equipment and computer readable storage medium for representing graph structure
CN111339762B (en) * 2020-02-14 2023-04-07 广州大学 Topic representation model construction method and device based on hybrid intelligence
CN111339783B (en) * 2020-02-24 2022-11-25 东南大学 RNTM-based topic mining method and device
CN111639189B (en) * 2020-04-29 2023-03-21 西北工业大学 Text graph construction method based on text content features
CN111738303B (en) * 2020-05-28 2023-05-23 华南理工大学 Long-tail distribution image recognition method based on hierarchical learning
CN111858931B (en) * 2020-07-08 2022-05-13 华中师范大学 Text generation method based on deep learning
CN111949790A (en) * 2020-07-20 2020-11-17 重庆邮电大学 Emotion classification method based on LDA topic model and hierarchical neural network
CN111966792B (en) * 2020-09-03 2023-07-25 网易(杭州)网络有限公司 Text processing method and device, electronic equipment and readable storage medium
CN112861982A (en) * 2021-02-24 2021-05-28 佛山市南海区广工大数控装备协同创新研究院 Long-tail target detection method based on gradient average
CN115563284B (en) * 2022-10-24 2023-06-23 重庆理工大学 Deep multi-instance weak supervision text classification method based on semantics

Citations (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN107092596A (en) * 2017-04-24 2017-08-25 重庆邮电大学 Text emotion analysis method based on attention CNNs and CCR
CN107168957A (en) * 2017-06-12 2017-09-15 云南大学 A kind of Chinese word cutting method
CN107368613A (en) * 2017-09-05 2017-11-21 中国科学院自动化研究所 Short text sentiment analysis method and device
WO2018085722A1 (en) * 2016-11-04 2018-05-11 Salesforce.Com, Inc. Quasi-recurrent neural network
CN108446275A (en) * 2018-03-21 2018-08-24 北京理工大学 Long text emotional orientation analytical method based on attention bilayer LSTM
CN108460019A (en) * 2018-02-28 2018-08-28 福州大学 A kind of emerging much-talked-about topic detecting system based on attention mechanism
CN108460013A (en) * 2018-01-30 2018-08-28 大连理工大学 A kind of sequence labelling model based on fine granularity vocabulary representation model

Family Cites Families (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US10268671B2 (en) * 2015-12-31 2019-04-23 Google Llc Generating parse trees of text segments using neural networks

Patent Citations (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2018085722A1 (en) * 2016-11-04 2018-05-11 Salesforce.Com, Inc. Quasi-recurrent neural network
CN107092596A (en) * 2017-04-24 2017-08-25 重庆邮电大学 Text emotion analysis method based on attention CNNs and CCR
CN107168957A (en) * 2017-06-12 2017-09-15 云南大学 A kind of Chinese word cutting method
CN107368613A (en) * 2017-09-05 2017-11-21 中国科学院自动化研究所 Short text sentiment analysis method and device
CN108460013A (en) * 2018-01-30 2018-08-28 大连理工大学 A kind of sequence labelling model based on fine granularity vocabulary representation model
CN108460019A (en) * 2018-02-28 2018-08-28 福州大学 A kind of emerging much-talked-about topic detecting system based on attention mechanism
CN108446275A (en) * 2018-03-21 2018-08-24 北京理工大学 Long text emotional orientation analytical method based on attention bilayer LSTM

Non-Patent Citations (5)

* Cited by examiner, † Cited by third party
Title
Entity recognition from clinical texts via recurrent neural network;Zengjian Liu等;《BMC MEDICAL INFORMATICS AND DECISION MAKING》;20170705;第53-60页 *
Representation learning for very short texts using weighted word embedding aggregation;Cedric De Boom等;《Pattern Recognition Letters》;20160901;第150-156页 *
一种基于深度学习与Labeled-LDA的文本分类方法;庞宇明;《中国优秀硕士学位论文全文数据库信息科技辑》;20180315;第I138-2153页 *
基于CSLSTM网络的文本情感分类;庄丽榕等;《计算机系统应用》;20180228;第230-235页 *
基于LSTM-Attention神经网络的文本特征提取方法;赵勤鲁等;《现代电子技术》;20180514;第167-170页 *

Also Published As

Publication number Publication date
CN109241377A (en) 2019-01-18

Similar Documents

Publication Publication Date Title
CN109241377B (en) Text document representation method and device based on deep learning topic information enhancement
CN111737474B (en) Method and device for training business model and determining text classification category
Wang et al. Adversarial cross-modal retrieval
Devika et al. Sentiment analysis: a comparative study on different approaches
CN110008338B (en) E-commerce evaluation emotion analysis method integrating GAN and transfer learning
CN111291556B (en) Chinese entity relation extraction method based on character and word feature fusion of entity meaning item
CN110059217A (en) A kind of image text cross-media retrieval method of two-level network
Ganai et al. Predicting next word using RNN and LSTM cells: Stastical language modeling
CN109101490B (en) Factual implicit emotion recognition method and system based on fusion feature representation
Huang et al. Siamese network-based supervised topic modeling
CN111753088A (en) Method for processing natural language information
Tran et al. Aggregating image and text quantized correlated components
KR20160149050A (en) Apparatus and method for selecting a pure play company by using text mining
CN108470025A (en) Partial-Topic probability generates regularization own coding text and is embedded in representation method
CN112256904A (en) Image retrieval method based on visual description sentences
CN115131613A (en) Small sample image classification method based on multidirectional knowledge migration
CN111259147A (en) Sentence-level emotion prediction method and system based on adaptive attention mechanism
Goswami et al. Classification of printed Gujarati characters using SOM based k-Nearest Neighbor Classifier
Zhang et al. Character confidence based on N-best list for keyword spotting in online Chinese handwritten documents
Wei et al. Study of text classification methods for data sets with huge features
CN113177120B (en) Quick information reorganizing method based on Chinese text classification
CN116151258A (en) Text disambiguation method, electronic device and storage medium
Kim et al. CNN based sentence classification with semantic features using word clustering
CN115357715A (en) Short text clustering method based on singular value decomposition and field pre-training
Preetham et al. Comparative Analysis of Research Papers Categorization using LDA and NMF Approaches

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant