CN109918501A - Method, apparatus, equipment and the storage medium of news article classification - Google Patents
Method, apparatus, equipment and the storage medium of news article classification Download PDFInfo
- Publication number
- CN109918501A CN109918501A CN201910046633.9A CN201910046633A CN109918501A CN 109918501 A CN109918501 A CN 109918501A CN 201910046633 A CN201910046633 A CN 201910046633A CN 109918501 A CN109918501 A CN 109918501A
- Authority
- CN
- China
- Prior art keywords
- model
- training
- language
- language model
- parameter
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Pending
Links
Landscapes
- Information Retrieval, Db Structures And Fs Structures Therefor (AREA)
Abstract
This application involves artificial intelligence fields, provide the method, apparatus and storage medium of a kind of news article classification, which comprises obtain the first data set, pre-process to first data set, obtain training set and test set;The parameter of the disaggregated model is obtained using training set pre-training disaggregated model using layering characteristic manner;The parameter in the disaggregated model is adjusted, after obtaining the optimal model parameters of the disaggregated model, constructs language model;The test set is tested using the language model, if it is detected that the accuracy rate of the language model is higher than preset threshold, it is determined that the language model meets class condition;Classified according to the language model to the second data set for inputting the language model.It provides and uses this programme, can be improved the accuracy rate of news article classification, and improve classifying quality of the term vector model to news article of word2vec pre-training.
Description
Technical field
This application involves method, apparatus, equipment and storages that artificial intelligence field more particularly to a kind of news article are classified
Medium.
Background technique
When classifying to news article, frequently with term vector model (word to vector, Word2vec) pre-training
Word is embedded in vector by term vector model, which is included in the first layer of neural network, and the rest part of neural network remains unchanged
Re -training is wanted, so being detached from the relationship of context when will lead to classification.The classifying quality for eventually leading to term vector model is poor.
Summary of the invention
This application provides method, apparatus, equipment and the storage mediums of a kind of classification of news article, are able to solve existing skill
The poor problem of the accuracy rate and effect that news article is classified in art.
In a first aspect, the application provides a kind of method of news article classification, which comprises
The first data set is obtained, first data set may include more news articles;
First data set is pre-processed, training set and test set are obtained;
The parameter of the disaggregated model is obtained using training set pre-training disaggregated model using layering characteristic manner;
The parameter in the disaggregated model is adjusted, after obtaining the optimal model parameters of the disaggregated model, constructs language mould
Type;
The test set is tested using the language model, if it is detected that the accuracy rate of the language model is higher than
Preset threshold, it is determined that the language model meets class condition;
Classified according to the language model to the second data set for inputting the language model.
It is described that institute is obtained using training set pre-training disaggregated model using layering characteristic manner in a kind of possible design
State the parameter of disaggregated model, comprising:
For the general levels structure of feature in the training set, from edge to shape, learn all levels from down to height
Feature, the ability to express of the inner link and language construction of continuous text in the training set is extracted, with the training classification
Model;
Wherein, the parameter of the disaggregated model indicates the weight of neural network, using the parameter of the disaggregated model as defeated
The vectorization for entering the word of language model indicates.
In a kind of possible design, parameter in the adjustment disaggregated model obtains the optimal of the disaggregated model
Model parameter, comprising:
In simulation space search, matches one and the training set is most matched it is assumed that obtain one group of optimal models
Parameter.
It is described assuming that space search, matches one and the most matched vacation of the training set in a kind of possible design
If to obtain one group of optimal model parameters, comprising:
The training set is inputted into the simulation space;
The disaggregated model is trained using the training set in the simulation space, training obtain one group with it is described
The most matched optimal model parameters of training set.
In a kind of possible design, the parameter in the language model includes classification independent variable and classification dependent variable, described
Construct language model, comprising:
Classification independent variable is set by headline and news author, sets classification dependent variable for news category;
According to the headline and the news author, the news category and optimal model parameters building
The language model.
In a kind of possible design, the disaggregated model includes ELMo model, OPenAI GPT model or Bert model.
In a kind of possible design, the pretreatment includes stratified sampling, the missing values processing in data and feature comb
Reason and screening.
Second aspect, the application provide a kind of device for classifying to news article, have and realize and correspond to above-mentioned the
On the one hand a kind of function of the method for news article classification provided.The function can also be passed through by hardware realization
Hardware executes corresponding software realization.Hardware or software include one or more modules corresponding with above-mentioned function, the mould
Block can be software and/or hardware.
In a kind of possible design, described device includes:
Input/output module, for obtaining the first data set, first data set may include more news articles;
Processing module obtains training set and test set for pre-processing to first data set;Using layered sheet
Sign mode obtains the parameter of the disaggregated model using training set pre-training disaggregated model;Ginseng in the whole disaggregated model
Number after obtaining the optimal model parameters of the disaggregated model, constructs language model;Using the language model to the test set
It is tested, if it is detected that the accuracy rate of the language model is higher than preset threshold, it is determined that the language model meets classification
Condition;Classified according to the second data set that the language model inputs the language model to the input/output module.
In a kind of possible design, the processing module is specifically used for:
For the general levels structure of feature in the training set, from edge to shape, learn all levels from down to height
Feature, the ability to express of the inner link and language construction of continuous text in the training set is extracted, with the training classification
Model;
Wherein, the parameter of the disaggregated model indicates the weight of neural network, using the parameter of the disaggregated model as defeated
The vectorization for entering the word of language model indicates.
In a kind of possible design, the processing module is specifically used for:
In simulation space search, matches one and the training set is most matched it is assumed that obtain one group of optimal models
Parameter.
In a kind of possible design, the processing module is specifically used for:
The training set is inputted into the simulation space by the input/output module;
The disaggregated model is trained using the training set in the simulation space, training obtain one group with it is described
The most matched optimal model parameters of training set.
In a kind of possible design, the parameter in the language model, including classification independent variable and classification dependent variable, it is described
Processing module is specifically used for:
Classification independent variable is set by headline and news author, sets classification dependent variable for news category;
According to the headline and the news author, the news category and optimal model parameters building
The language model.
In a kind of possible design, the disaggregated model includes ELMo model, OPenAI GPT model or Bert model.Institute
Stating pretreatment includes stratified sampling, the missing values processing in data and feature combing and screening.
The another aspect of the application provides a kind of computer equipment comprising processor, the memory of at least one connection
And input-output unit, wherein the memory is for storing program code, and the processor is for calling in the memory
Program code execute method described in above-mentioned first aspect.
The another aspect of the application provides a kind of computer storage medium comprising instruction, when it runs on computers
When, so that computer executes method described in above-mentioned first aspect.
Compared to the prior art, in scheme provided by the present application, first data set is pre-processed, is trained
Collection and test set;The parameter of the disaggregated model is obtained using training set pre-training disaggregated model using layering characteristic manner;
The parameter in the disaggregated model is adjusted, after obtaining the optimal model parameters of the disaggregated model, constructs language model;Using institute
It states language model to test the test set, if it is detected that the accuracy rate of the language model is higher than preset threshold, really
The fixed language model meets class condition;The second data set for inputting the language model is carried out according to the language model
Classification.There is provided use this programme, can be improved news article classification accuracy rate, and improve word2vec pre-training word to
Model is measured to the classifying quality of news article.
Detailed description of the invention
Fig. 1 is a kind of flow diagram of method legal person of news article classification in the embodiment of the present application;
Fig. 2 is a kind of structural schematic diagram of the device in the embodiment of the present application for classifying to news article;
Fig. 3 is another structural schematic diagram of the device in the embodiment of the present application for classifying to news article.
The embodiments will be further described with reference to the accompanying drawings for realization, functional characteristics and the advantage of the application purpose.
Specific embodiment
It should be appreciated that specific embodiment described herein is not used to limit the application only to explain the application.This
The specification and claims of application and term " first " in above-mentioned attached drawing, " second " etc. are for distinguishing similar right
As without being used to describe a particular order or precedence order.It should be understood that the data used in this way in the appropriate case can be with
It exchanges, so that the embodiments described herein can be implemented with the sequence other than the content for illustrating or describing herein.In addition,
Term " includes " and " having " and their any deformation, it is intended that cover it is non-exclusive include, for example, containing a system
The process, method, system, product or equipment of column step or module those of are not necessarily limited to be clearly listed step or module, and
Being may include other steps or module being not clearly listed or intrinsic for these process, methods, product or equipment, this
The division of module appeared in application, only a kind of division in logic can have other when realizing in practical application
Division mode, such as multiple modules can be combined into or are integrated in another system, or some features can be ignored, or not held
Row.
The application provides method, apparatus, equipment and the storage medium of a kind of news article classification, can be used for news category.
In order to solve the above technical problems, the application it is main the following technical schemes are provided:
The news article of acquisition is pre-processed, constructs pre-training language model (for example, by using ELMo model, OPenAI
GPT model or Bert model), pre-training language model can be handled super large text or super large expectation, be made full use of big
Single language corpus of scale.Classified by pre-training language model to these news articles, is able to solve above-mentioned background technique
In technical problem, improve news article classification accuracy rate.
Fig. 1 is please referred to, a kind of method for providing news article classification to the application below is illustrated, the method
Include:
101, the first data set is obtained.
Wherein, first data set may include more news articles.First data set owner will include news article
Article title, article abstract, article summarize etc. information.The news article of each news platform can be obtained in a manner of crawler.
102, first data set is pre-processed, obtains training set and test set.
Wherein, training set is used to construct language model, and test set is used to examine the accuracy rate of the language model of building
It tests.
In some embodiments, the pretreatment includes stratified sampling, missing values processing and feature combing in data
With screening.
Wherein, stratified sampling, which refers to, carries out stochastical sampling to each classification respectively, to guarantee in sample space or type
Uniformity and representativeness in selection.For example, divided according to story label (such as according to labels such as tourism, amusement, social activities
Divide news article).
Missing values processing in data includes extraction, cleaning, conversion, integrated and filling.
Denoising mode can be used with screening in feature combing.
In some embodiments, pretreatment can use and reserve method or cross-validation method.Wherein, method is reserved to refer to institute
The set that the first data set is divided into two mutual exclusions is stated, that is, is divided into test set and training set.Cross-validation method refers to will be described
First data set is divided into the similar exclusive subsets of k size, and each subset keeps the consistency of data distribution as far as possible, i.e.,
Each subset is concentrated through stratified sampling from the first data and obtains.Then, use the union of k-1 subset as training every time
Collection, remaining subset is as test set;It is obtained with k group training/test set, in this way so as to carry out k training and survey
Examination, what is finally returned that is the mean value of k test result.
103, the parameter of the disaggregated model is obtained using training set pre-training disaggregated model using layering characteristic manner.
Wherein, the parameter of disaggregated model can become the weight of neural network, using the parameter of disaggregated model as input language
Say that the vectorization of the word of model indicates, i.e. term vector.Term vector is used to measure the similitude between word and word.
By taking word2vec model as an example, the application carrys out the entire word2vec model of pre-training using layering characteristic manner, throws
It has abandoned the mode of the first layer of in word2vec model initialization model in current mechanism.Such as study word vector ratio
Make to learn image border in computer vision, then general levels structure one of the layering characteristic manner just as learning characteristics of image
Sample from edge to shape, then arrives high-level semantics concept.When carrying out the entire word2vec model of pre-training using layering characteristic manner,
These pre-training also acquire the feature of low-level and the spy of high-level while carrying out pre-training to entire word2vec model
Sign enables word2vec model to learn into text higher level nuance, and the general of image is arrived in study
Feature.
In some embodiments, disaggregated model can choose ELMo model, OPenAI GPT (Generative Pre-
) or the sorter models such as Bert model Training.
Wherein, ELMo is a kind of novel depth contextualized vocabulary sign, can carry out complex characteristic (such as syntax and semantic) to word
It is modeled and (polysemant is modeled) with variation of the word in language context.The term vector of the application is the two-way language of depth
The function for saying model (biLM) internal state, pre-training forms in one large-scale text corpus.Specifically, one is first trained
A complete language model, then go processing to need the text of training with this language model, generate corresponding term vector, ELMo mould
Type can generate different term vectors to the same word in different sentences.After this good language model of pre-training, ELMo is just
It is to be used as word expression according to formula, is in fact exactly that each middle layer of this bi-directional language model is carried out a summation.Most
Simply top expression can be used also as ELMo.It then, can be by ELMo in the NLP task for carrying out having supervision
Directly as in the term vector input either top expression of model of merging features to specific tasks model.
ELMo be when learning language model gone from entire corpus study, by language model generate word to
Amount is equivalent to the term vector learnt based on entire corpus, therefore, is capable of the meaning of more one word of accurate representation.
Bert model is that (i.e. pre-training language indicates for a kind of multi-layer biaxially oriented Transformer encoder based on fine tuning
Method), input indicates to indicate single text sentence or a pair of of text in a word sequence, for example, word sequence is expressed as
[problem, answer].Bert model makes that general " language understanding " can be trained on large-scale text corpus (such as wikipedia)
Model.
104, the parameter in the disaggregated model is adjusted, after obtaining the optimal model parameters of the disaggregated model, constructs language
Say model.
In some embodiments, the parameter adjusted in the disaggregated model obtains the optimal mould of the disaggregated model
Shape parameter, comprising:
In simulation space search, matches one and the training set is most matched it is assumed that obtain one group of optimal models
Parameter.It simulates space and alternatively referred to as assumes space.
For example, the training set is inputted the simulation space, in the simulation space using the training set to described
Disaggregated model is trained, and final training obtains one group and the most matched optimal model parameters of the training set.
105, the test set is tested using the language model, if it is detected that the accuracy rate of the language model
Higher than preset threshold, it is determined that the language model meets class condition.
Class condition can include: at least one text feature is same or similar, wherein text feature can be news article
The labels such as label, such as finance and economics, amusement, sport, science and technology, military affairs or household.Text feature can also be that area name is (military
The Chinese, Shenzhen etc.), the offer platform (such as newspaper, radio station, broadcast, internet) of news article.It can also be regional rank (example
Such as city-level, provincial or area grade).
106, classified according to the language model to the second data set for inputting the language model.
Compared with current mechanism, in the embodiment of the present application, first data set is pre-processed, obtain training set and
Test set;The parameter of the disaggregated model is obtained using training set pre-training disaggregated model using layering characteristic manner;Adjustment
Parameter in the disaggregated model after obtaining the optimal model parameters of the disaggregated model, constructs language model;Utilize institute's predicate
Speech model tests the test set, if it is detected that the accuracy rate of the language model is higher than preset threshold, it is determined that institute
It states language model and meets class condition;The second data set for inputting the language model is divided according to the language model
Class.It provides and uses this programme, can be improved the accuracy rate of news article classification, and improve the term vector of word2vec pre-training
Classifying quality of the model to news article.
Optionally, described using layering characteristic manner in some embodiments of the present application, utilize training set pre-training point
Class model obtains the parameter of the disaggregated model, comprising:
For the general levels structure of feature in the training set, from edge to shape, learn all levels from down to height
Feature, the ability to express of the inner link and language construction of continuous text in the training set is extracted, with the training classification
Model;
Wherein, the parameter of the disaggregated model indicates the weight of neural network, using the parameter of the disaggregated model as defeated
The vectorization for entering the word of language model indicates.
Optionally, in some embodiments of the present application, the parameter in the language model includes classification independent variable and divides
Class dependent variable, the building language model, comprising:
Classification independent variable is set by headline and news author, sets classification dependent variable for news category;
According to the headline and the news author, the news category and optimal model parameters building
The language model.
Every technical characteristic mentioned in above-described embodiment is applied equally to corresponding to Fig. 2 and Fig. 3 in the application
Embodiment, subsequent similar place repeats no more.
The method of news article a kind of in the application classification is illustrated above, below to execution above-mentioned news article point
The device of the method for class is described.
A kind of structural schematic diagram of device 20 for classifying to news article as shown in Figure 2, can be applied to news
Article classification.Device 20 in the embodiment of the present application can be realized corresponding to performed in embodiment corresponding to above-mentioned Fig. 1
Step in the method for news article classification.The function that device 20 is realized can also be held by hardware realization by hardware
The corresponding software realization of row.Hardware or software include one or more modules corresponding with above-mentioned function, and the module can be with
It is software and/or hardware.Described device 20 may include input/output module 201 and processing module 202, the processing module 202
Realizing with the function of input/output module 201 can refer to operation performed in embodiment corresponding to Fig. 1, not go to live in the household of one's in-laws on getting married herein
It states.Processing module can be used for controlling the input for obtaining module 201 or output operation.
In some embodiments, the input/output module 201 can be used for obtaining the first data set, first data set
It may include more news articles;
The processing module 202 can be used for pre-processing first data set, obtain training set and test set;It adopts
The parameter of the disaggregated model is obtained using training set pre-training disaggregated model with layering characteristic manner;The whole disaggregated model
In parameter, after obtaining the optimal model parameters of the disaggregated model, construct language model;Using the language model to described
Test set is tested, if it is detected that the accuracy rate of the language model is higher than preset threshold, it is determined that the language model symbol
Close class condition;Classified according to the language model to the second data set for inputting the language model.
In the embodiment of the present application, processing module 202 pre-processes first data set, obtains training set and test
Collection;The parameter of the disaggregated model is obtained using training set pre-training disaggregated model using layering characteristic manner;Described in adjustment
Parameter in disaggregated model after obtaining the optimal model parameters of the disaggregated model, constructs language model;Utilize the language mould
Type tests the test set, if it is detected that the accuracy rate of the language model is higher than preset threshold, it is determined that institute's predicate
Speech model meets class condition;Classified according to the language model to the second data set for inputting the language model.It mentions
For using this programme, the accuracy rate of news article classification can be improved, and improve the term vector model of word2vec pre-training
To the classifying quality of news article.
In some embodiments, the processing module 202 is specifically used for:
For the general levels structure of feature in the training set, from edge to shape, learn all levels from down to height
Feature, the ability to express of the inner link and language construction of continuous text in the training set is extracted, with the training classification
Model;
Wherein, the parameter of the disaggregated model indicates the weight of neural network, using the parameter of the disaggregated model as defeated
The vectorization for entering the word of language model indicates.
In some embodiments, the processing module 202 is specifically used for:
In simulation space search, matches one and the training set is most matched it is assumed that obtain one group of optimal models
Parameter.
In some embodiments, the processing module 202 is specifically used for:
The training set is inputted into the simulation space by the input/output module 201;
The disaggregated model is trained using the training set in the simulation space, training obtain one group with it is described
The most matched optimal model parameters of training set.
In some embodiments, parameter in the language model, including classification independent variable and classification dependent variable, the place
Reason module 202 is specifically used for:
Classification independent variable is set by headline and news author, sets classification dependent variable for news category;
According to the headline and the news author, the news category and optimal model parameters building
The language model.
In some embodiments, the disaggregated model includes ELMo model, OPenAI GPT model or Bert model.It is described
Pretreatment includes stratified sampling, the missing values processing in data and feature combing and screening.
It describes in the embodiment of the present application and is used for news article point respectively from the angle of modular functionality entity above
The device of class introduces a kind of computer equipment from hardware point of view below, as shown in figure 3, comprising: processor, memory, input
Output unit and storage are in the memory and the computer program that can run on the processor.For example, the calculating
Machine program can be the corresponding program of method of news article classification in embodiment corresponding to Fig. 1.For example, working as computer equipment
When realizing the function of device 20 as shown in Figure 2, the processor is realized corresponding to above-mentioned Fig. 2 when executing the computer program
Embodiment in each step in the method for news article classification that is executed by device 20;Alternatively, described in the processor execution
The function of each module in the device 20 of embodiment corresponding to above-mentioned Fig. 2 is realized when computer program.In another example the computer journey
Sequence can be the corresponding program of method of news article classification in embodiment corresponding to Fig. 1.
Alleged processor can be central processing unit (Central Processing Unit, CPU), can also be it
His general processor, digital signal processor (Digital Signal Processor, DSP), specific integrated circuit
(Application Specific Integrated Circuit, ASIC), ready-made programmable gate array (Field-
Programmable Gate Array, FPGA) either other programmable logic device, discrete gate or transistor logic,
Discrete hardware components etc..General processor can be microprocessor or the processor is also possible to any conventional processor
Deng the processor is the control centre of the computer equipment, utilizes various interfaces and the entire computer equipment of connection
Various pieces.
The memory can be used for storing the computer program and/or module, and the processor is by operation or executes
Computer program in the memory and/or module are stored, and calls the data being stored in memory, described in realization
The various functions of computer equipment.The memory can mainly include storing program area and storage data area, wherein storage program
It area can application program (such as sound-playing function, image player function etc.) needed for storage program area, at least one function
Deng;Storage data area, which can be stored, uses created data (such as audio data, video data etc.) etc. according to mobile phone.This
Outside, memory may include high-speed random access memory, can also include nonvolatile memory, such as hard disk, memory, insert
Connect formula hard disk, intelligent memory card (Smart Media Card, SMC), secure digital (Secure Digital, SD) card, flash memory
Block (Flash Card), at least one disk memory, flush memory device or other volatile solid-state parts.
The input-output unit can also be replaced with input unit and output unit, can be same or different object
Manage entity.When for identical physical entity, input-output unit may be collectively referred to as.The input-output unit can be transceiver.
The memory can integrate in the processor, can also be provided separately with the processor.
Through the above description of the embodiments, those skilled in the art can be understood that above-described embodiment side
Method can be realized by means of software and necessary general hardware platform, naturally it is also possible to by hardware, but in many cases
The former is more preferably embodiment.Based on this understanding, the technical solution of the application substantially in other words does the prior art
The part contributed out can be embodied in the form of software products, which is stored in a storage medium
In (such as ROM/RAM), including some instructions are used so that a terminal (can be mobile phone, computer, server or network are set
It is standby etc.) execute method described in each embodiment of the application.
Embodiments herein is described above in conjunction with attached drawing, but the application be not limited to it is above-mentioned specific
Embodiment, the above mentioned embodiment is only schematical, rather than restrictive, those skilled in the art
Under the enlightenment of the application, when not departing from the application objective and scope of the claimed protection, can also it make very much
Form, it is all using equivalent structure or equivalent flow shift made by present specification and accompanying drawing content, directly or indirectly
Other related technical areas are used in, these are belonged within the protection of the application.
Claims (10)
1. a kind of method of news article classification, which is characterized in that the described method includes:
The first data set is obtained, first data set may include more news articles;
First data set is pre-processed, training set and test set are obtained;
The parameter of the disaggregated model is obtained using training set pre-training disaggregated model using layering characteristic manner;
The parameter in the disaggregated model is adjusted, after obtaining the optimal model parameters of the disaggregated model, constructs language model;
The test set is tested using the language model, is preset if it is detected that the accuracy rate of the language model is higher than
Threshold value, it is determined that the language model meets class condition;
Classified according to the language model to the second data set for inputting the language model.
2. being instructed in advance the method according to claim 1, wherein described use layering characteristic manner using training set
Practice disaggregated model, obtain the parameter of the disaggregated model, comprising:
For the general levels structure of feature in the training set, from edge to shape, from the spy down to high all levels of study
Sign, extracts the ability to express of the inner link and language construction of continuous text in the training set, with the training disaggregated model;
Wherein, the parameter of the disaggregated model indicates the weight of neural network, using the parameter of the disaggregated model as input language
Say that the vectorization of the word of model indicates.
3. method according to claim 1 or 2, which is characterized in that the parameter in the adjustment disaggregated model obtains
The optimal model parameters of the disaggregated model, comprising:
In simulation space search, matches one and the training set is most matched it is assumed that obtain one group of optimal model parameters.
4. according to the method described in claim 3, it is characterized in that, it is described assuming that space search, match one with it is described
Training set is most matched it is assumed that obtain one group of optimal model parameters, comprising:
The training set is inputted into the simulation space;
The disaggregated model is trained using the training set in the simulation space, training obtains one group and the training
Collect most matched optimal model parameters.
5. according to the method described in claim 4, it is characterized in that, parameter in the language model include classification independent variable and
Classification dependent variable, the building language model, comprising:
Classification independent variable is set by headline and news author, sets classification dependent variable for news category;
According to the headline and the news author, the news category and optimal model parameters building
Language model.
6. the method according to claim 1, wherein the disaggregated model includes ELMo model, OPenAI GPT
Model or Bert model.
7. the method according to claim 1, wherein the pretreatment includes the missing in stratified sampling, data
Value processing and feature combing and screening.
8. a kind of device for classifying to news article, which is characterized in that described device includes:
Input/output module, for obtaining the first data set, first data set may include more news articles;
Processing module obtains training set and test set for pre-processing to first data set;Using layering characterization side
Formula obtains the parameter of the disaggregated model using training set pre-training disaggregated model;Parameter in the whole disaggregated model, obtains
To after the optimal model parameters of the disaggregated model, language model is constructed;The test set is carried out using the language model
Test, if it is detected that the accuracy rate of the language model is higher than preset threshold, it is determined that the language model meets class condition;
Classified according to the second data set that the language model inputs the language model to the input/output module.
9. a kind of computer equipment, which is characterized in that described device includes:
At least one processor, memory and input-output unit;
Wherein, the memory is for storing program code, and the processor is for calling the program stored in the memory
Code is executed such as method of any of claims 1-7.
10. a kind of computer storage medium, which is characterized in that it includes instruction, when run on a computer, so that calculating
Machine executes such as method of any of claims 1-7.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201910046633.9A CN109918501A (en) | 2019-01-18 | 2019-01-18 | Method, apparatus, equipment and the storage medium of news article classification |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201910046633.9A CN109918501A (en) | 2019-01-18 | 2019-01-18 | Method, apparatus, equipment and the storage medium of news article classification |
Publications (1)
Publication Number | Publication Date |
---|---|
CN109918501A true CN109918501A (en) | 2019-06-21 |
Family
ID=66960321
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN201910046633.9A Pending CN109918501A (en) | 2019-01-18 | 2019-01-18 | Method, apparatus, equipment and the storage medium of news article classification |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN109918501A (en) |
Cited By (10)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN110347840A (en) * | 2019-07-18 | 2019-10-18 | 携程计算机技术(上海)有限公司 | Complain prediction technique, system, equipment and the storage medium of text categories |
CN110390110A (en) * | 2019-07-30 | 2019-10-29 | 阿里巴巴集团控股有限公司 | The method and apparatus that pre-training for semantic matches generates sentence vector |
CN110930022A (en) * | 2019-11-20 | 2020-03-27 | 携程计算机技术(上海)有限公司 | Hotel static information detection method and system, electronic equipment and storage medium |
CN110991535A (en) * | 2019-12-04 | 2020-04-10 | 中山大学 | pCR prediction method based on multi-type medical data |
CN111221974A (en) * | 2020-04-22 | 2020-06-02 | 成都索贝数码科技股份有限公司 | Method for constructing news text classification model based on hierarchical structure multi-label system |
CN111930947A (en) * | 2020-08-26 | 2020-11-13 | 施建军 | System and method for identifying authors of modern Chinese written works |
CN111966828A (en) * | 2020-07-27 | 2020-11-20 | 电子科技大学 | Newspaper and magazine news classification method based on text context structure and attribute information superposition network |
CN112632283A (en) * | 2020-12-30 | 2021-04-09 | 北京有竹居网络技术有限公司 | Model generation method, text classification method, device, equipment and medium |
CN113722493A (en) * | 2021-09-09 | 2021-11-30 | 北京百度网讯科技有限公司 | Data processing method, device, storage medium and program product for text classification |
CN115496062A (en) * | 2022-11-10 | 2022-12-20 | 杭州费尔斯通科技有限公司 | Method, system, computer device and storage medium for recognizing enterprise address selection willingness |
Citations (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN107025284A (en) * | 2017-04-06 | 2017-08-08 | 中南大学 | The recognition methods of network comment text emotion tendency and convolutional neural networks model |
CN107301246A (en) * | 2017-07-14 | 2017-10-27 | 河北工业大学 | Chinese Text Categorization based on ultra-deep convolutional neural networks structural model |
CN108363753A (en) * | 2018-01-30 | 2018-08-03 | 南京邮电大学 | Comment text sentiment classification model is trained and sensibility classification method, device and equipment |
-
2019
- 2019-01-18 CN CN201910046633.9A patent/CN109918501A/en active Pending
Patent Citations (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN107025284A (en) * | 2017-04-06 | 2017-08-08 | 中南大学 | The recognition methods of network comment text emotion tendency and convolutional neural networks model |
CN107301246A (en) * | 2017-07-14 | 2017-10-27 | 河北工业大学 | Chinese Text Categorization based on ultra-deep convolutional neural networks structural model |
CN108363753A (en) * | 2018-01-30 | 2018-08-03 | 南京邮电大学 | Comment text sentiment classification model is trained and sensibility classification method, device and equipment |
Non-Patent Citations (2)
Title |
---|
SEBASTIAN RUDER: "NLP\'s ImageNet moment has arrived", pages 1 - 20, Retrieved from the Internet <URL:https://www.ruder.io/nlp-imagenet/> * |
雷峰网: "ImageNet带来的预训练模型之风,马上要吹进NLP领域了", pages 1 - 9, Retrieved from the Internet <URL:https://www.leiphone.com/category/ai/5hn7YLV31DEREEeA.html> * |
Cited By (15)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN110347840A (en) * | 2019-07-18 | 2019-10-18 | 携程计算机技术(上海)有限公司 | Complain prediction technique, system, equipment and the storage medium of text categories |
CN110390110A (en) * | 2019-07-30 | 2019-10-29 | 阿里巴巴集团控股有限公司 | The method and apparatus that pre-training for semantic matches generates sentence vector |
CN110390110B (en) * | 2019-07-30 | 2023-06-27 | 创新先进技术有限公司 | Method and apparatus for pre-training generation of sentence vectors for semantic matching |
CN110930022A (en) * | 2019-11-20 | 2020-03-27 | 携程计算机技术(上海)有限公司 | Hotel static information detection method and system, electronic equipment and storage medium |
CN110991535A (en) * | 2019-12-04 | 2020-04-10 | 中山大学 | pCR prediction method based on multi-type medical data |
CN111221974A (en) * | 2020-04-22 | 2020-06-02 | 成都索贝数码科技股份有限公司 | Method for constructing news text classification model based on hierarchical structure multi-label system |
CN111221974B (en) * | 2020-04-22 | 2020-08-14 | 成都索贝数码科技股份有限公司 | Method for constructing news text classification model based on hierarchical structure multi-label system |
CN111966828B (en) * | 2020-07-27 | 2022-05-03 | 电子科技大学 | Newspaper and magazine news classification method based on text context structure and attribute information superposition network |
CN111966828A (en) * | 2020-07-27 | 2020-11-20 | 电子科技大学 | Newspaper and magazine news classification method based on text context structure and attribute information superposition network |
CN111930947A (en) * | 2020-08-26 | 2020-11-13 | 施建军 | System and method for identifying authors of modern Chinese written works |
CN112632283A (en) * | 2020-12-30 | 2021-04-09 | 北京有竹居网络技术有限公司 | Model generation method, text classification method, device, equipment and medium |
CN113722493A (en) * | 2021-09-09 | 2021-11-30 | 北京百度网讯科技有限公司 | Data processing method, device, storage medium and program product for text classification |
CN113722493B (en) * | 2021-09-09 | 2023-10-13 | 北京百度网讯科技有限公司 | Text classification data processing method, apparatus and storage medium |
CN115496062A (en) * | 2022-11-10 | 2022-12-20 | 杭州费尔斯通科技有限公司 | Method, system, computer device and storage medium for recognizing enterprise address selection willingness |
CN115496062B (en) * | 2022-11-10 | 2023-02-28 | 杭州费尔斯通科技有限公司 | Method and system for identifying enterprise address selection willingness, computer equipment and storage medium |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN109918501A (en) | Method, apparatus, equipment and the storage medium of news article classification | |
Lines et al. | Time series classification with HIVE-COTE: The hierarchical vote collective of transformation-based ensembles | |
CN112632385B (en) | Course recommendation method, course recommendation device, computer equipment and medium | |
Zhang et al. | MOOCRC: A highly accurate resource recommendation model for use in MOOC environments | |
CN110263324A (en) | Text handling method, model training method and device | |
CN111160350B (en) | Portrait segmentation method, model training method, device, medium and electronic equipment | |
CN111741330B (en) | Video content evaluation method and device, storage medium and computer equipment | |
CN106126492B (en) | Sentence recognition methods and device based on two-way LSTM neural network | |
Wen et al. | Dynamic interactive multiview memory network for emotion recognition in conversation | |
Hong et al. | Sentiment analysis with deeply learned distributed representations of variable length texts | |
CN108960407A (en) | Recurrent neural network language model training method, device, equipment and medium | |
CN110852360A (en) | Image emotion recognition method, device, equipment and storage medium | |
CN110825850B (en) | Natural language theme classification method and device | |
CN108416032A (en) | A kind of file classification method, device and storage medium | |
CN106502979A (en) | A kind of data processing method of natural language information and device | |
CN109933782A (en) | User emotion prediction technique and device | |
CN109284370A (en) | A kind of mobile application description and permission fidelity determination method and device based on deep learning | |
CN113392640B (en) | Title determination method, device, equipment and storage medium | |
CN114154570A (en) | Sample screening method and system and neural network model training method | |
Jia et al. | A multimodal emotion recognition model integrating speech, video and MoCAP | |
CN117390497B (en) | Category prediction method, device and equipment based on large language model | |
CN112948710A (en) | Graph neural network-based punishment education recommendation method, system and storage medium | |
CN108021565A (en) | A kind of analysis method and device of the user satisfaction based on linguistic level | |
CN116304063B (en) | Simple emotion knowledge enhancement prompt tuning aspect-level emotion classification method | |
CN116757195A (en) | Implicit emotion recognition method based on prompt learning |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination |