CN111930952A - Method, system, equipment and storage medium for long text cascade classification - Google Patents
Method, system, equipment and storage medium for long text cascade classification Download PDFInfo
- Publication number
- CN111930952A CN111930952A CN202010991960.4A CN202010991960A CN111930952A CN 111930952 A CN111930952 A CN 111930952A CN 202010991960 A CN202010991960 A CN 202010991960A CN 111930952 A CN111930952 A CN 111930952A
- Authority
- CN
- China
- Prior art keywords
- interval
- long text
- vector
- semantic
- text
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Pending
Links
- 238000000034 method Methods 0.000 title claims abstract description 34
- 238000003860 storage Methods 0.000 title claims description 7
- 239000013598 vector Substances 0.000 claims abstract description 82
- 230000007246 mechanism Effects 0.000 claims abstract description 14
- 238000007781 pre-processing Methods 0.000 claims abstract description 13
- 238000009826 distribution Methods 0.000 claims abstract description 12
- 238000013145 classification model Methods 0.000 claims abstract description 9
- 230000009467 reduction Effects 0.000 claims abstract description 8
- 238000004590 computer program Methods 0.000 claims description 11
- 230000003993 interaction Effects 0.000 claims description 8
- 230000011218 segmentation Effects 0.000 claims description 8
- 238000004364 calculation method Methods 0.000 claims description 6
- 238000012549 training Methods 0.000 claims description 5
- 238000004422 calculation algorithm Methods 0.000 claims description 4
- 238000005520 cutting process Methods 0.000 claims description 4
- 230000008569 process Effects 0.000 claims description 4
- 239000011159 matrix material Substances 0.000 claims description 3
- 238000005070 sampling Methods 0.000 claims description 3
- 238000005303 weighing Methods 0.000 claims description 2
- 238000011161 development Methods 0.000 abstract description 4
- 238000000605 extraction Methods 0.000 description 4
- 230000018109 developmental process Effects 0.000 description 3
- 238000005516 engineering process Methods 0.000 description 3
- 208000027418 Wounds and injury Diseases 0.000 description 2
- 238000004458 analytical method Methods 0.000 description 2
- 230000007547 defect Effects 0.000 description 2
- 238000010586 diagram Methods 0.000 description 2
- 238000011144 upstream manufacturing Methods 0.000 description 2
- 235000006508 Nelumbo nucifera Nutrition 0.000 description 1
- 240000002853 Nelumbo nucifera Species 0.000 description 1
- 238000013473 artificial intelligence Methods 0.000 description 1
- 230000009286 beneficial effect Effects 0.000 description 1
- 230000006378 damage Effects 0.000 description 1
- 230000000694 effects Effects 0.000 description 1
- 230000006870 function Effects 0.000 description 1
- 230000006872 improvement Effects 0.000 description 1
- 208000014674 injury Diseases 0.000 description 1
- 238000012986 modification Methods 0.000 description 1
- 230000004048 modification Effects 0.000 description 1
- 238000003058 natural language processing Methods 0.000 description 1
- 230000000630 rising effect Effects 0.000 description 1
- 238000006467 substitution reaction Methods 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F16/00—Information retrieval; Database structures therefor; File system structures therefor
- G06F16/30—Information retrieval; Database structures therefor; File system structures therefor of unstructured textual data
- G06F16/35—Clustering; Classification
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F40/00—Handling natural language data
- G06F40/30—Semantic analysis
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Physics & Mathematics (AREA)
- General Engineering & Computer Science (AREA)
- General Physics & Mathematics (AREA)
- Audiology, Speech & Language Pathology (AREA)
- Computational Linguistics (AREA)
- General Health & Medical Sciences (AREA)
- Health & Medical Sciences (AREA)
- Artificial Intelligence (AREA)
- Data Mining & Analysis (AREA)
- Databases & Information Systems (AREA)
- Information Retrieval, Db Structures And Fs Structures Therefor (AREA)
Abstract
The invention relates to a long text cascade classification method, which comprises the following steps: s1: preprocessing input long text data by using a sliding window mechanism, and dividing the long text data into a plurality of intervals; s2: semantic coding is carried out on the generated texts in each interval, and a local semantic vector of each interval is obtained; s3: extracting keywords from the interval text, coding the keywords into keyword vectors, and respectively performing vector splicing on the keyword vectors and the local semantic vectors to obtain an overall semantic vector of the long text; s4: reducing the dimension of the whole semantic vector, and calculating the probability distribution of class labels of the whole semantic vector after dimension reduction by using a classifier; s5: according to the steps S1-S4, a classification model of the long text corpus is trained. The technical scheme provided by the invention can improve the performance of the classification model at the bottom layer to promote the development of other intelligent services in the vertical field, thereby improving the experience and viscosity of users on intelligent products.
Description
Technical Field
The invention belongs to the technical field of natural language processing, and particularly relates to a method, a system, equipment and a storage medium for long text cascade classification.
Background
With the continuous popularization of artificial intelligence technology, the industry development in the vertical field is changing day by day, and novel AI products are continuously rising.
Products such as case retrieval, regulation retrieval, intelligent consulting robot and the like in the vertical field of law all need support of intelligent semantic analysis technology, but the intelligent semantic analysis technology comprises a plurality of algorithm models such as an element extraction model, a relation extraction model and a complex event extraction model in the field of law, and the models are supported by using a classification model as a bottom layer. The performance of the bottom layer is closely related to the application performance of the upper layer model, however, part of the open corpus in the vertical field such as the legal field is in the form of long text, taking the referee document as an example, and the minimum document length is between 500 words and 1000 words.
For the long text, the existing methods in the vertical field shorten the text length by three modes of interception, segmentation and summarization, and the methods generally have 2 defects:
(1) each term within a document is computed interactively with all other terms, resulting in high complexity, especially if the length of the document exceeds a threshold.
(2) Different section texts obtained by segmentation cannot be subjected to cross-section interaction, and a large amount of key information can be lost.
The two defects greatly limit the performance of the existing classification model in the vertical field and the application range of the upper layer model.
Disclosure of Invention
In order to solve the above problems, the present invention provides a method, a system, a device and a storage medium for long text cascade classification, which can improve the performance of a bottom-layer classification model to promote the development of other intelligent services in the vertical field, and further improve the experience and viscosity of users on intelligent products.
The technical scheme of the invention is as follows:
a long text cascade classification method comprises the following steps:
s1: preprocessing input long text data by using a sliding window mechanism, and dividing the long text data into a plurality of intervals;
s2: semantic coding is carried out on the generated texts in each interval, and a local semantic vector of each interval is obtained;
s3: extracting keywords from the interval text, coding the keywords into keyword vectors, and respectively performing vector splicing on the keyword vectors and the local semantic vectors to obtain an overall semantic vector of the long text;
s4: reducing the dimension of the whole semantic vector, and calculating the probability distribution of class labels of the whole semantic vector after dimension reduction by using a classifier;
s5: according to the steps S1-S4, a classification model of the long text corpus is trained.
Preferably, the preprocessing in step S1 includes the specific steps of:
s1.1: the long text is segmented according to the length d of the specified interval to obtain a first interval s1;
S2.2: performing subsequent interval segmentation according to the specified interval length d and the step length overlap on the basis of the step S1.1 to obtain an interval S of the residual text2-skThe number of the segments is totally divided into k segments.
Preferably, the specific step of acquiring the local semantic vector in step S2 is:
s2.1: performing token preprocessing on the texts in all the obtained intervals;
s2.2: performing semantic coding on a token set of each interval obtained after token preprocessing through an encoder to obtain a corresponding interval semantic vector;
s2.3: and extracting corresponding interval global features aiming at the interval semantic vectors to form local semantic vectors of the whole long text on each interval.
Preferably, the encoder is a BERT encoder, in a BERT mechanism,' [ CLS]' the corresponding vector includes the global feature of the section, the section global feature corresponding to the CLS is extracted in the step S2.3, and the local semantic vector of the whole long text on each section is formed。
Preferably, the specific process of acquiring the whole semantic vector in step S3 is as follows:
s3.1: extracting topic keywords corresponding to each interval of the long text by using an LDA algorithm based on Gibbs sampling, and obtaining a keyword vector theta corresponding to the topic keywords through an encoder;
s3.2: splicing and summing the topic keyword vectors corresponding to the intervals according to the following formula:
j is the number of the keyword sets extracted from the ith interval;
and splicing the local semantic vectors on the interval according to the following formula:
s3.3: an attention mechanism is introduced to enhance the dependence of the long text on the key word features of the topic in each interval, so as to enhance the overall semantic expression of the classification features, and the specific calculation formula is as follows:
wherein, wi、biRepresenting the weight matrix and the bias during training; u. ofiFor the weighing of the ith interval s as a scoring functioniThe importance of the topic keyword feature of (1); a isiIs a vector s in the intervaliCalculating the obtained weight value through a softmax formula; f is the final interval siAnd the corresponding attention weighted vector is the whole semantic vector of the long text.
Preferably, the specific calculation process of the dimensionality reduction and the probability distribution in step S4 is as follows:
s4.1: carrying out concentration and dimension reduction on the whole semantic vector f of the long text, wherein the formula is as follows:
converting the whole semantic vector f of the long text into a low-dimensional condensed vector f2;
S4.2: for low dimensional concentration vector f2Calculating the probability distribution of the class label by utilizing softmax, wherein the formula is as follows:
the invention provides a long text cascade classification system, which utilizes a long text cascade classification method based on a sliding window mechanism and is characterized by comprising the following steps:
the sliding window module: used for cutting the input long text to form a plurality of sections;
the coding module: the local semantic coding is carried out on the text in each interval to obtain a local semantic vector;
a global interaction module: acquiring an integral semantic vector of the long text according to the local semantic vector and the text in each interval;
a classification module: and the method is used for classifying the whole semantic vector to obtain the probability distribution of the class labels.
The invention also provides a computer device comprising a memory, a processor and a computer program stored in the memory and executable on the processor, wherein the processor implements the steps of the long text cascade classification method when executing the computer program.
The invention also provides a computer-readable storage medium, in which a computer program is stored, which computer program, when being executed by a processor, is adapted to carry out the steps of the long text cascade classification method.
The invention has the beneficial effects that:
1. according to the technical scheme, the development of other intelligent services in the vertical field is promoted by improving the performance of the classification model at the bottom layer, so that the experience and viscosity of a user on an intelligent product are improved;
2. according to the technical scheme, an attention mechanism is introduced to enhance the dependence of the long text on the key word features of the section topics, and further the overall semantic expression of the classification features is enhanced;
3. the classification module provided by the invention performs long text corpus training through the sliding window module, the coding module, the global interaction module and the classification module, improves the capturing capability of the model on the local semantic features and the global dependency features of the long text, and further promotes the performance improvement of upstream tasks.
Drawings
Fig. 1 is an overall architecture diagram of a method provided in an embodiment of the invention.
Fig. 2 is a diagram illustrating a sliding window module according to an embodiment of the present invention.
FIG. 3 is a detailed flowchart of keyword extraction in the global interaction module according to the present invention.
Detailed Description
The embodiments of the present invention will be described in detail below with reference to the accompanying drawings.
As shown in fig. 1, an embodiment of the present invention provides a long text cascade classification method, including the following steps:
1. and preprocessing the input long text data.
1) And cutting for the first time.
The long text is segmented according to the length d of the specified interval to obtain a first interval s1Taking the legal field as an example, as shown in fig. 2, the first interval obtained is: [ Zhangxiu lotus root cause family.... The rights of liabilities]。
2) And continuously cutting.
And performing subsequent interval segmentation on the input text on the basis of the first segmentation according to the specified interval length d and the step length overlap to obtain all intervals.
Taking the legal field as an example, as shown in fig. 2, the second interval s is obtained after the segmentation is continued2Is given as [ warranty]Wherein s is2Is the first word "has the position of and s1Is overlap between the positions of the first word "sheet".
2. And (4) carrying out semantic coding on the texts in all the intervals obtained in the step (1).
1) And (5) token pretreatment.
And carrying out token preprocessing on the texts in all the obtained intervals. Since BERT is an encoder using bi-directional encoding and has a global semantic association feature, BERT mechanism is used herein for illustration, and other BERT mechanism-like models can be used.
Taking the legal field as an example, for the text token preprocessing, the text token becomes [ ' [ CLS ] ', ' Zhang ', ' Xiu ', ' Juan ', ' because of family trivia ', ' because ' the token operation, the text token becomes [ ', ' family ', ' trivia ', ' President ', ' Code ', ' AND ', ' West ', ' Fu ', ' alarm ', ' leaving ', ' wedding ', ' SEP ] ' ] after the token operation.
2) And independently coding interval semantics.
And carrying out semantic coding on each interval token set obtained after token preprocessing by a BERT (belief transform) coder to obtain a corresponding interval semantic vector. For example, the intervals S1 and S2 are coded by BERT to obtain semantic vectors。
3) And extracting local semantics.
Due to the BERT mechanism,' [ CLS ]]' the corresponding vector includes the global feature of the interval, so the semantic vector V is aimed at each interval1、V2、......、VkExtracting the interval global features corresponding to the CLS to form the local semantic direction of the whole long text on each intervalMeasurement of。
3. And acquiring the integral semantic vector of the long text through the local semantic vector.
1) And extracting the key words.
And extracting the topic key words corresponding to each interval of the long text by using an LDA algorithm based on Gibbs sampling. For example, fig. 2 illustrates that the topic keywords extracted from the first interval of the legal long text are [ 'family trivia', 'guarantee' ]. And then obtaining a keyword vector theta corresponding to the topic keywords through a BERT encoder.
2) And vector splicing.
Splicing and summing the topic keyword vectors corresponding to the intervals according to the following formula:
wherein j is the number of the keyword sets extracted from the ith interval.
Splicing the interval vectors according to the following formula:
3) attention interaction calculation.
Not all words have the same effect on the expression of the semantic information of the sentence, and similarly, the influence of texts in each interval on the expression of the overall semantics of the long text is different. Therefore, an attention mechanism is introduced to enhance the dependence of the long text on the key word features of the topic in each interval, and further enhance the overall semantic expression of the classification features. The specific calculation method is as follows:
wherein, wi、biRepresenting the weight matrix and the bias during training; u. ofiAnalogous to a scoring function, for measuring the ith interval siThe importance of the topic keyword feature of (1); a isiI.e. the interval vector siCalculating the obtained weight value through a softmax formula; f is the final interval siA corresponding attention weighting vector.
4. And performing dimensionality reduction and concentration on the whole semantic vector of the long text, and acquiring a corresponding category label.
1) And two layers are fully connected.
After obtaining the whole semantic vector f of the long text, concentrating and reducing dimensions of the long text, and operating according to the following formula:
by w as in the above two formulaef1、wf2Converting f into a low-dimensional concentration vector f2。
2) Softmax calculates the class probability distribution.
For a global vector f as follows2Calculating the probability distribution of the class label:
taking the legal field as an example, if the labels corresponding to the legal long text are 4 types of divorce disputes, divorce property disputes, labor and wound disputes and house property trade disputes, the probability distribution p available for the example text of fig. 2 is { 'divorce dispute': 0.2, 'divorce property disputes': 0.6, 'labor and industrial injury dispute': 0.1, 'house property trade dispute': 0.1}. According to the maximum probability principle, the finally obtained prediction category of the example text is divorce property disputes.
5. The model training is completed according to the method provided by the first 4 steps to obtain the classification model of the long text corpus, so that the capturing capability of the model on the local semantic features and the global dependency features of the long text corpus can be improved, and the performance of the upstream task is promoted by the method.
The method is not limited to be used in the legal field, and other fields including long texts, such as medical treatment, finance and the like, can be applied.
Based on the long text cascade classification method, the embodiment of the invention also provides a long text cascade classification system, which comprises a sliding window module for performing sliding window segmentation on the input long text, a coding module for performing local semantic coding on the sliding window text, a global interaction module for interacting global information, and a classification module for classifying the global information, as shown in fig. 1.
The sliding window module cuts the text into a plurality of interval texts by using a sliding window mechanism aiming at the input long text;
the coding module codes the local semantics corresponding to the segmented interval text to obtain interval semantic features; the global interaction module is used for integrating global attention aiming at the interval semantic features corresponding to the intervals so as to obtain global features; the classification module classifies the global features to obtain prediction categories.
The invention also provides a computer device comprising a memory, a processor and a computer program stored in the memory and executable on the processor, wherein the processor implements the steps of the long text cascade classification method when executing the computer program.
The invention also provides a computer-readable storage medium, in which a computer program is stored, which computer program, when being executed by a processor, is adapted to carry out the steps of the long text cascade classification method.
Finally, it should be noted that: the above-mentioned embodiments are only specific embodiments of the present invention, which are used for illustrating the technical solutions of the present invention and not for limiting the same, and the protection scope of the present invention is not limited thereto, although the present invention is described in detail with reference to the foregoing embodiments, those skilled in the art should understand that: any person skilled in the art can modify or easily conceive the technical solutions described in the foregoing embodiments or equivalent substitutes for some technical features within the technical scope of the present disclosure; such modifications, changes or substitutions do not depart from the spirit and scope of the present invention in its spirit and scope. Are intended to be covered by the scope of the present invention. Therefore, the protection scope of the present invention shall be subject to the protection scope of the claims.
Claims (9)
1. A long text cascade classification method is characterized by comprising the following steps:
s1: preprocessing input long text data by using a sliding window mechanism, and dividing the long text data into a plurality of intervals;
s2: semantic coding is carried out on the generated texts in each interval, and a local semantic vector of each interval is obtained;
s3: extracting keywords from the interval text, coding the keywords into keyword vectors, and respectively performing vector splicing on the keyword vectors and the local semantic vectors to obtain an overall semantic vector of the long text;
s4: reducing the dimension of the whole semantic vector, and calculating the probability distribution of class labels of the whole semantic vector after dimension reduction by using a classifier;
s5: according to the steps S1-S4, a classification model of the long text corpus is trained.
2. The method for classifying long text cascades of claim 1, wherein the specific steps of preprocessing in step S1 are:
s1.1: the long text is segmented according to the length d of the specified interval to obtain a first interval s1;
S1.2: performing subsequent interval segmentation according to the specified interval length d and the step length overlap on the basis of the step S1.1 to obtain an interval S of the residual text2-skThe number of the segments is totally divided into k segments.
3. The method for classifying long texts in cascade according to claim 1, wherein the step S2 of obtaining local semantic vectors includes the following specific steps:
s2.1: performing token preprocessing on the texts in all the obtained intervals;
s2.2: performing semantic coding on a token set of each interval obtained after token preprocessing through an encoder to obtain a corresponding interval semantic vector;
s2.3: and extracting corresponding interval global features aiming at the interval semantic vectors to form local semantic vectors of the whole long text on each interval.
4. The method of claim 3, wherein the encoder is a BERT encoder,' [ CLS ] in BERT mechanism]' the corresponding vector includes the global feature of the section, the section global feature corresponding to the CLS is extracted in the step S2.3, and the local semantic vector of the whole long text on each section is formed。
5. The method for classifying long text cascades of claim 1, wherein the specific process of obtaining the whole semantic vector in step S3 is as follows:
s3.1: extracting topic keywords corresponding to each interval of the long text by using an LDA algorithm based on Gibbs sampling, and obtaining a keyword vector theta corresponding to the topic keywords through an encoder;
s3.2: splicing and summing the topic keyword vectors corresponding to the intervals according to the following formula:
j is the number of the keyword sets extracted from the ith interval;
and splicing the local semantic vectors on the interval according to the following formula:
s3.3: an attention mechanism is introduced to enhance the dependence of the long text on the key word features of the topic in each interval, so as to enhance the overall semantic expression of the classification features, and the specific calculation formula is as follows:
wherein, wi、biRepresenting the weight matrix and the bias during training; u. ofiFor the weighing of the ith interval s as a scoring functioniThe importance of the topic keyword feature of (1); a isiIs a vector s in the intervaliCalculating the obtained weight value through a softmax formula; f is the final interval siAnd the corresponding attention weighted vector is the whole semantic vector of the long text.
6. The method for classifying long text cascades according to claim 1, wherein the specific calculation processes of dimension reduction and probability distribution in the step S4 are as follows:
s4.1: carrying out concentration and dimension reduction on the whole semantic vector f of the long text, wherein the formula is as follows:
converting the whole semantic vector f of the long text into a low-dimensional condensed vector f2;
S4.2: for low dimensional concentration vector f2And calculating the probability distribution of the class label by utilizing softmax, wherein the formula is as follows:
7. a long text cascade classification system using the method for classifying long text based on the sliding window mechanism of any one of claims 1 to 6, comprising:
the sliding window module: used for cutting the input long text to form a plurality of sections;
the coding module: the local semantic coding is carried out on the text in each interval to obtain a local semantic vector;
a global interaction module: acquiring an integral semantic vector of the long text according to the local semantic vector and the text in each interval;
a classification module: and the method is used for classifying the whole semantic vector to obtain the probability distribution of the class labels.
8. A computer device comprising a memory, a processor and a computer program stored in the memory and executable on the processor, characterized in that the steps of the method according to any of claims 1 to 6 are implemented when the computer program is executed by the processor.
9. A computer-readable storage medium, in which a computer program is stored which, when being executed by a processor, carries out the steps of the method according to any one of claims 1 to 6.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202010991960.4A CN111930952A (en) | 2020-09-21 | 2020-09-21 | Method, system, equipment and storage medium for long text cascade classification |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202010991960.4A CN111930952A (en) | 2020-09-21 | 2020-09-21 | Method, system, equipment and storage medium for long text cascade classification |
Publications (1)
Publication Number | Publication Date |
---|---|
CN111930952A true CN111930952A (en) | 2020-11-13 |
Family
ID=73334223
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN202010991960.4A Pending CN111930952A (en) | 2020-09-21 | 2020-09-21 | Method, system, equipment and storage medium for long text cascade classification |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN111930952A (en) |
Cited By (7)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN112507120A (en) * | 2021-02-07 | 2021-03-16 | 上海二三四五网络科技有限公司 | Prediction method and device for keeping classification consistency |
CN112699231A (en) * | 2020-12-25 | 2021-04-23 | 科讯嘉联信息技术有限公司 | Work order abstract summarizing method based on sliding window correlation calculation and Copy mechanism |
CN113378580A (en) * | 2021-06-23 | 2021-09-10 | 北京百度网讯科技有限公司 | Document layout analysis method, model training method, device and equipment |
CN113761931A (en) * | 2020-11-23 | 2021-12-07 | 北京沃东天骏信息技术有限公司 | Information processing method, device, electronic equipment and storage medium |
CN113935335A (en) * | 2021-12-16 | 2022-01-14 | 深圳市迪博企业风险管理技术有限公司 | Intelligent checking method for contract text compliance |
CN115630142A (en) * | 2022-12-08 | 2023-01-20 | 中国传媒大学 | Multi-language long text similarity retrieval and classification tool |
CN116663568A (en) * | 2023-07-31 | 2023-08-29 | 腾云创威信息科技(威海)有限公司 | Critical task identification system and method based on priority |
Citations (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN111177376A (en) * | 2019-12-17 | 2020-05-19 | 东华大学 | Chinese text classification method based on BERT and CNN hierarchical connection |
CN111401061A (en) * | 2020-03-19 | 2020-07-10 | 昆明理工大学 | Method for identifying news opinion involved in case based on BERT and Bi L STM-Attention |
CN111597340A (en) * | 2020-05-22 | 2020-08-28 | 迪爱斯信息技术股份有限公司 | Text classification method and device and readable storage medium |
-
2020
- 2020-09-21 CN CN202010991960.4A patent/CN111930952A/en active Pending
Patent Citations (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN111177376A (en) * | 2019-12-17 | 2020-05-19 | 东华大学 | Chinese text classification method based on BERT and CNN hierarchical connection |
CN111401061A (en) * | 2020-03-19 | 2020-07-10 | 昆明理工大学 | Method for identifying news opinion involved in case based on BERT and Bi L STM-Attention |
CN111597340A (en) * | 2020-05-22 | 2020-08-28 | 迪爱斯信息技术股份有限公司 | Text classification method and device and readable storage medium |
Non-Patent Citations (2)
Title |
---|
RAGHAVENDRA PAPPAGARI 等: "HIERARCHICAL TRANSFORMERS FOR LONG DOCUMENT CLASSIFICATION", 《HTTPS://ARXIV.ORG/ABS/1910.10781》 * |
SAURABH GOYAL 等: "PoWER-BERT: Accelerating BERT Inference via Progressive Word-vector Elimination", 《ICML 2020》 * |
Cited By (9)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN113761931A (en) * | 2020-11-23 | 2021-12-07 | 北京沃东天骏信息技术有限公司 | Information processing method, device, electronic equipment and storage medium |
CN112699231A (en) * | 2020-12-25 | 2021-04-23 | 科讯嘉联信息技术有限公司 | Work order abstract summarizing method based on sliding window correlation calculation and Copy mechanism |
CN112507120A (en) * | 2021-02-07 | 2021-03-16 | 上海二三四五网络科技有限公司 | Prediction method and device for keeping classification consistency |
CN113378580A (en) * | 2021-06-23 | 2021-09-10 | 北京百度网讯科技有限公司 | Document layout analysis method, model training method, device and equipment |
CN113935335A (en) * | 2021-12-16 | 2022-01-14 | 深圳市迪博企业风险管理技术有限公司 | Intelligent checking method for contract text compliance |
CN113935335B (en) * | 2021-12-16 | 2022-03-22 | 深圳市迪博企业风险管理技术有限公司 | Intelligent checking method for contract text compliance |
CN115630142A (en) * | 2022-12-08 | 2023-01-20 | 中国传媒大学 | Multi-language long text similarity retrieval and classification tool |
CN116663568A (en) * | 2023-07-31 | 2023-08-29 | 腾云创威信息科技(威海)有限公司 | Critical task identification system and method based on priority |
CN116663568B (en) * | 2023-07-31 | 2023-11-17 | 腾云创威信息科技(威海)有限公司 | Critical task identification system and method based on priority |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN111930952A (en) | Method, system, equipment and storage medium for long text cascade classification | |
CN109597997B (en) | Comment entity and aspect-level emotion classification method and device and model training thereof | |
Sivakumar et al. | Review on word2vec word embedding neural net | |
Gao et al. | Convolutional neural network based sentiment analysis using Adaboost combination | |
CN106844349B (en) | Comment spam recognition methods based on coorinated training | |
Zhang | Research on text classification method based on LSTM neural network model | |
CN113553510B (en) | Text information recommendation method and device and readable medium | |
CN113282714B (en) | Event detection method based on differential word vector representation | |
CN112667782A (en) | Text classification method, device, equipment and storage medium | |
CN115796182A (en) | Multi-modal named entity recognition method based on entity-level cross-modal interaction | |
CN111401062A (en) | Text risk identification method, device and equipment | |
Rajalakshmi et al. | Sentimental analysis of code-mixed Hindi language | |
Mutanga et al. | Detecting hate speech on Twitter network using ensemble machine learning | |
Mansour et al. | Text vectorization method based on concept mining using clustering techniques | |
Schwartz et al. | SoPa: Bridging CNNs, RNNs, and weighted finite-state machines | |
CN114490954A (en) | Document level generation type event extraction method based on task adjustment | |
Samir et al. | Twitter sentiment analysis using BERT | |
Kavatagi et al. | A context aware embedding for the detection of hate speech in social media networks | |
Derbentsev et al. | Sentiment analysis of electronic social media based on deep learning | |
Assiri et al. | DeBERTa-GRU: Sentiment Analysis for Large Language Model. | |
Igual et al. | Basics of Natural Language Processing | |
Bhowmik et al. | Sentiment analysis with hotel customer reviews using FNet | |
Shi et al. | Improve on Entity Recognition Method Based on BiLSTM-CRF Model for the Nuclear Technology Knowledge Graph | |
CN113076754A (en) | False comment detection method and system based on knowledge integration | |
Manna et al. | Application of deep learning techniques on document classification |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
RJ01 | Rejection of invention patent application after publication | ||
RJ01 | Rejection of invention patent application after publication |
Application publication date: 20201113 |