CN101621391A - Method and system for classifying short texts based on probability topic - Google Patents

Method and system for classifying short texts based on probability topic Download PDF

Info

Publication number
CN101621391A
CN101621391A CN200910090377A CN200910090377A CN101621391A CN 101621391 A CN101621391 A CN 101621391A CN 200910090377 A CN200910090377 A CN 200910090377A CN 200910090377 A CN200910090377 A CN 200910090377A CN 101621391 A CN101621391 A CN 101621391A
Authority
CN
China
Prior art keywords
short text
text
short
probability
module
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN200910090377A
Other languages
Chinese (zh)
Inventor
刘文印
权小军
张加龙
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
BEIJING BAIWENBAIDA NETWORK TECHNOLOGIES Co Ltd
Original Assignee
BEIJING BAIWENBAIDA NETWORK TECHNOLOGIES Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by BEIJING BAIWENBAIDA NETWORK TECHNOLOGIES Co Ltd filed Critical BEIJING BAIWENBAIDA NETWORK TECHNOLOGIES Co Ltd
Priority to CN200910090377A priority Critical patent/CN101621391A/en
Publication of CN101621391A publication Critical patent/CN101621391A/en
Pending legal-status Critical Current

Links

Images

Landscapes

  • Information Retrieval, Db Structures And Fs Structures Therefor (AREA)

Abstract

The invention discloses a method and a system for classifying short texts based on the probability topic; the method is used in a data processing device of a question answering system for classifying the short texts according to the similarity of the short texts. The method comprises the following steps: respectively acquiring initialized text vectors according to an input target short text and a short text acquired from the database of the question answering system; scanning the two short texts to respectively acquire differentiating words of the two short texts; when the relevance degrees of the differentiating words of the two short texts and a probability topic are higher than a threshold, modifying the text vectors of the two short texts according to the relevance degrees; working out the similarity of the two short texts according to the modified text vectors of the two short texts; acquiring another short text form the database of the question answering system till all short texts in the database are traversed; and then executing a scanning step; classifying the target short text according to the similarity.

Description

Carry out the method and system of short text classification based on probability topic
Technical field
The present invention relates to utilize data mining to carry out the areas of information technology of text-processing, particularly relate to a kind of method and system that carries out the short text classification based on probability topic, and the short text search method, the spam recognition methods.
Background technology
In today of information technology develop rapidly, the user can obtain bulk information by multiple channel.For example by browsing page, utilize search engine to retrieve, receive note, mail, operational network question answering system etc.But the problem that occurs is data magnanimity, poor information usually.
For example, occur a large amount of mails in the mailbox, wherein both comprised normal work mail or life mail, comprise spam again.Comprise a large amount of useless advertisement SMSs in the note that receives.In the network question answering system, when the user proposes a problem, may obtain a large amount of answers from other users.Possible some departed from theme, some answer very accurate.When in webpage, retrieving a keyword, obtain a large amount of web page interlinkages that comprises this keyword.Wherein some web page interlinkage can embody its corresponding webpage and possesses the content relevant with this keyword, and the correlation of some web page interlinkage is relatively poor.A large amount of models in the forum are the time sequencing free arrangement, independent separately, the arrangement of jumping of content.
Because how gigantic project when being reluctant maybe can not traverse every data, obtains active data from mass data, the perhaps data of potentially useful, or be to get rid of unnecessary data are problem demanding prompt solutions as the user.
Current, mass data occurs with the form of short text.The answer of for example above-mentioned note, mail title, web search, network question answering system etc.If can realize automatic classification based on its inner link (for example semantic relation) is carried out in the answer of note, mail title, web page interlinkage, network question answering system, by the user categories of interest (content) is read one by one again, perhaps, identification shielding rubbish information will significantly reduce the time that the user is spent when locating valid data.Generally speaking, dwindle the scope that the user need check one by one, that can significantly improve the user consults convenience and operating efficiency, improves user experience.
For achieving the above object, at first need short text is carried out data mining, particularly carry out the differentiation of classification according to its inherent meaning relevance.
In the prior art, exist by calculating similarity between the text and carry out the method for class discrimination.It is by representing that with its word frequency vector (or being called text word frequency vector, term-frequency vector) distance of calculating then between the word frequency vector obtains text similarity to text.Yet the method for most of existing calculating text similarities only is applicable to long text, and why effective traditional calculating long text similarity based method is, is because similar long text has comprised the identical vocabulary of some usually.But for short text, similar short text might not have identical speech, and the flexibility of natural language makes people to express the identical meaning by different words, so existing similarity calculating method effect and bad.
Proposed some solutions at this problem in the prior art, its thought generally is by means of a dictionary (as WordNet) short text to be expanded.In general, this method can be found the correlation between speech and the speech more accurately, and still, many times the true relation between speech and the speech depends on concrete application background.Just concrete semantic environment.
How at the few characteristics of vocabulary identical between short text, and according to the contact of semanteme, mark off different classes of, and particular category extracted or exclude, realizing the high speed location valid data in the information services such as short message service, search service, automatic question answering service, mail service, is to need the problem that solves in the present industry.
Summary of the invention
The problem that the present invention solves is, at information service high speeds such as short message service, search service, automatic question answering service, mail service location valid data, short text is classified based on its inherent semantic relation, and then extract target data efficiently based on this classification.
The invention discloses a kind of method of carrying out short text classification, be applied in the data processing equipment of question answering system, be used for short text being classified, comprising according to the similarity between short text (should be) with being problem or answer based on probability topic:
According to a target short text and a short text that is obtained from the database of question answering system of input, obtain its initialized text vector respectively;
Scan this two short text, obtain the differentiating words of this two short text respectively;
When the correlation degree of the differentiating words of this two short text and a probability theme all is higher than threshold value, according to the text vector of this this two short text of correlation degree correction;
Calculate similarity between two short texts according to the text vector of revised this two short text;
Obtain another short text in the database of question answering system, the short text in database is finished by traversal, carries out this scanning step;
According to this similarity target short text is classified.
The invention also discloses a kind of system that carries out short text classification based on probability topic, be applied in the data processing equipment of question answering system, be used for short text being classified, comprising according to the similarity between short text (should be) with being problem or answer:
Initialization module, a target short text and a short text that is obtained from the database of question answering system according to input obtain its initialized text vector respectively;
The differentiating words identification module is used to scan this two short text, and obtains the differentiating words of this two short text respectively;
Judgment amendment module is when the correlation degree of the differentiating words of this two short text and a probability theme all is higher than threshold value, according to the text vector of this this two short text of correlation degree correction;
Computing module calculates similarity between two short texts according to the text vector of revised two short texts;
Sort module is classified to target short text according to similarity.
The invention also discloses a kind of method of carrying out the short text retrieval, be applied to comprise the steps: in the search engine server based on probability topic
Step 1 receives the retrieval short text as term that transmits from the search subscriber terminal;
Step 2 is obtained a short text from the database of search engine server;
Step 3 according to above-mentioned retrieval short text and the short text that obtains, is obtained its initialized text vector respectively from database;
Step 4 scans this two short text, obtains the differentiating words of this two short text respectively;
Step 5 is when the correlation degree of the differentiating words of this two short text and a probability theme all is higher than threshold value, according to the text vector of this this two short text of correlation degree correction;
Step 6 is calculated similarity between two short texts according to the text vector of revised this two short text;
Step 7, circulation execution in step 2, the short text traversal in database finishes;
Step 8 is extracted and the short text of the similarity of retrieving short text greater than predetermined value, is sent to the search subscriber terminal and shows.
The invention also discloses a kind of method of the identification spam based on probability topic, be applied to comprise the steps: in the mail server
Step 1, storage model spam short text in the database of mail server, and obtain one of them;
Step 2 is obtained a short text from the database of mail server;
Step 3 according to above-mentioned two short texts, is obtained its initialized text vector respectively;
Step 4 scans this two short text, obtains the differentiating words of this two short text respectively;
Step 5 is when the correlation degree of the differentiating words of this two short text and a probability theme all is higher than threshold value, according to the text vector of this this two short text of correlation degree correction;
Step 6 is calculated similarity between two short texts according to the text vector of revised this two short text;
Step 7, the circulation execution in step 2 whole short text traversals in database finish, and after the whole short text traversals in database finished, execution in step 1 all traveled through until model spam short text and finishes again;
Step 8 will be designated spam greater than the short text that obtains of predetermined value with the similarity of this model spam short text from the database of mail server.
The invention also discloses a kind of method based on shielding rubbish, advertising message in the portable terminal of probability topic, be applied in the fire compartment wall of mobile communication system, store the model short text in the fire compartment wall, mobile communication system comprises short message server, fire compartment wall, user terminal, comprises the steps:
Step 1, fire compartment wall receive the short message text that this short message server sends;
Step 2, fire compartment wall are obtained a model short text;
Step 3 according to above-mentioned this short message text and this model short text, is obtained its initialized text vector respectively;
Step 4 scans this two short text, obtains the differentiating words of this two short text respectively;
Step 5 is when the correlation degree of the differentiating words of this two short text and a probability theme all is higher than threshold value, according to the text vector of this this two short text of correlation degree correction;
Step 6 is calculated similarity between two short texts according to the text vector of revised this two short text;
Step 7, circulation execution in step 2, all the model short text traversals in fire compartment wall finish;
Step 8, the similarity of fire compartment wall interception and model short text is higher than this short message text of a predetermined value.
The effect that the present invention realizes is that in question answering system, the problem at the user proposes according to its inherent meaning, is included into the classification space of a whole page accurately in the location automatically with it, makes that the logic correspondence of the problem that each classification space of a whole page is included with it is strong.The field that the answer user of being convenient to answer a question is good at according to self selects the space of a whole page to answer, and the problem that makes obtains very fast authority's response.Simultaneously, the inventive method has improved the accuracy that the short text similarity is calculated, and can give the expert who is familiar with this problem domain with question recommending more accurately.Make question answering system as the platform of information interchange, make the distribution of information more accurate.
In the short text search field, search method of the present invention is based on probability topic and retrieves, and whether reaches threshold value to judge its correlation in terms of content by judging similarity, has therefore also improved the accuracy of short text search return results greatly.
In getting rid of spam, shielding advertising message field, but effective recognition discharging garbage mail, advertising message, and indicated, shield.Improve the efficient and the accuracy of spam, advertising message identification.
Description of drawings
Figure 1A, 1B, 1C, 1D are depicted as a kind of structural representation that carries out the system of short text classification based on probability topic;
Fig. 2 A, 2B are for carrying out the flow chart of short text classification based on probability topic;
Fig. 3 is for carrying out the flow chart of question recommending based on probability topic;
Fig. 4 A is the structural representation of search engine server;
Fig. 4 B is for carrying out the flow chart of short text retrieval based on probability topic;
Fig. 5 A is the structural representation of mail server;
Fig. 5 B is for discerning the flow chart of spam based on probability topic;
Fig. 6 A is the structural representation of the present invention's one Application Example;
Fig. 6 B is the schematic flow sheet of the present invention's one Application Example;
Fig. 7 A, 7B are the little average and grand mean value schematic diagram of getting the F1 that the different themes number obtains on two data sets;
Fig. 8 A, 8B are depicted as the little average and grand mean value schematic diagram of getting the F1 that different λ values obtain on two data sets.
Embodiment
The invention discloses a kind of method and system of carrying out short text classification, can find out true relation between word and the word, and then calculate the similarity between the short text according to probability topic based on probability topic.With localizing objects data efficiently.
In the data processing equipment of the fire compartment wall of a mail server, search engine server, mobile communication system or question answering system server end, a kind of system that carries out the short text classification based on probability topic is set.
With the question answering system is example, and question answering system is an online interaction system, also is question answering system, realizes the computer processing system of user interactions question and answer.And to please refer to number of patent application be 200510130778.5 Chinese patent.
The technical problem that the present invention solves in question answering system is, the problem that proposes at the user according to its inherent meaning, is included into accurately that the classification space of a whole page shows in the location automatically with it, makes that the logic correspondence of the problem that each classification space of a whole page is included with it is strong.The field that the answer user of being convenient to answer a question is good at according to self selects the space of a whole page to answer, and the problem that makes obtains very fast authority's response.
See also shown in Figure 1A, the 1B, be a kind of structural representation that carries out the system of short text classification based on probability topic of the present invention.
The system 100 that carries out the short text classification based on probability topic comprises question answering system server end 10 and a plurality of terminal 20.Comprise a communication connection interface 11, database 12, data processing equipment 13, display module 14, input module 15 in the question answering system server end 10.
Communication connection interface 11 is used to receive the data from a plurality of terminals 20, for example target short text such as the problem that sends by terminal 20 of user or answer.Data processing equipment 13 is used for target short text that receives by communication connection interface 11 and the short text that extracts from database 12 are successively carried out analyzing and processing.Input module 15 is used for input parameter or supervisory instruction.Display module 14 is used for the short text after handling is by analysis shown.
Further be provided with a plurality of modules in this data processing equipment 13.Comprise:
Initialization module 1301 at a target short text that receives from communication connection interface 11 and the short text that obtains, and obtains its initialized text vector respectively from database 12;
Differentiating words identification module 1302 is used to scan this two short text, and obtains the differentiating words of this two short text respectively; Wherein, differentiating words be in a short text, occur and in another short text absent variable word.Differentiating words can hint out two relations between the short text usually.
Judgment amendment module 1303 is when the correlation degree of the differentiating words of this two short text and a probability theme all is higher than threshold value, according to the text vector of this this two short text of correlation degree correction;
Probability topic abstraction module 1304, the generating probability theme is used to sample;
Computing module 1305 calculates similarity between two short texts according to the text vector of revised this two short text;
Sort module 1306 is classified to target short text according to this similarity.
This sort module 1306 can be carried out concrete classification by the mode that adopts k nearest neighbour classification device.K nearest neighbour classification device is a known technology, its basic operation principle is, calculate target short text and be present in the similarity of the short text in the classification with all, select the highest K of a similarity text, the classification according to this K short text place determines which classification is target short text assign in the most at last then.
The specific implementation method is, calculates the correlation of each classification (some classifications at K short text place) and this target short text, and target short text is assigned in that classification of correlation maximum the most at last.Wherein, the computational methods of the correlation of classification and target short text are as follows, for example, set K=5, promptly choose 5 with the highest short text of target short text similarity, suppose short text d1, d2 belongs to classification C1, short text d3, d4, d5 belong to classification C2.The similarity of target short text and d1 represents with S1, and the similarity of target short text and d2 represents with S2, by that analogy.If S1+S2>S3+S4+S5 assigns to target short text among the classification C1 so, otherwise assigns among the C2.Also optional other numerical value of K.
Except that k nearest neighbour classification device, also can adopt additive method realization in the prior art, as centroid method.
The barycenter ratio juris is that each classification is expressed as a vector, calculates the similarity of target short text and each categorization vector, finds out the classification of that classification of similarity maximum as target short text.The method that each classification is expressed as vector is, the vector of all short texts in each classification is weighted on average, obtains an average vector as categorization vector.
Described differentiating words identification module 1302 further comprises: contrast module 1 and differentiating words set acquisition module 2.This contrast module 1 is used for this two short text is compared; This differentiating words set acquisition module 2 is used for obtaining respectively the differentiating words set of each short text.
Described judgment amendment module 1303 further comprises:
Module 3 chosen in theme, is used to choose a probability topic t i, search for respectively in the differentiating words set of two short texts at this probability topic t iThe speech of last probability maximum;
Judge module 4 is used to judge that selected two differentiating words are at probability topic t iOn probability whether all greater than threshold value, if all greater than, carry out correcting module 5; Otherwise, carry out circulation module 6;
Correcting module 5 is used for initialized text vector is revised.
Circulation module 6 is used to judge whether also have theme not to be selected, if, carry out theme and choose module 3, if not, carry out computing module 1305.
Introduce technical scheme of the present invention in detail by the following examples with reference to the above-mentioned system that carries out the short text classification based on probability topic, and in conjunction with idiographic flow.What see also Fig. 2 A carries out the flow chart that short text is classified based on probability topic.
Store a large amount of short text data that comprise problem, answer in the database 12, constitute a text set D.Text set D is made up of M short text, comprises N word that differs from one another, and promptly text set D has made up dictionary W={w 1, w 2..., w N.And text set D comprises Z theme T={t 1, t 2... t Z.d iRepresent a text among the D.
Terminal 20 transmission one target short text d 1Be sent to question answering system server end 10.
Step 201, the content in data processing equipment 13 these databases 12 of scanning utilizes initialization module 1301 to obtain one at random or in proper order not by traversal short text d 2, and to these two short text d 1, d 2Carry out the initialization of text vector respectively.
Its text vector of difference initialization V (1), V (2)For:
V ( 1 ) = { v 1 ( 1 ) , v 2 ( 1 ) , · · · , v N ( 1 ) }
V ( 2 ) = { v 1 ( 2 ) , v 2 ( 2 ) , · · · , v N ( 2 ) } ,
Wherein, v i (j)Be the w in the dictionary iAt short text d jIn weight (if do not occur this w in the short text i, v so i (j)Be zero); v i (j)Calculate with the TF-IDF method.TF-IDF is the state of the art.This kind initial method only is a kind of embodiment, carries out initialization with other similarity methods and is also included within the scope of the present invention.
In the TF-IDF method, v i (j)Can calculate by following formula:
v i ( j ) = tf ij × log ( M / df i ) , Wherein, M is the short-and-medium number of texts of text set D; Df jBe to comprise speech w among the text set D iThe number of short text; Tf IjBe w iAt short text d jThe middle number of times that occurs.
Additive method such as TF method are promptly only calculated tf IjPerhaps simple 0-1 assignment method, if promptly comprise i word in the short text, v so i (j)Be 1; Otherwise be 0.
Because the speech number that short text comprises is limited, some noise speech (for example, the speech that all occurs in many texts) can influence the short text calculation of similarity degree, therefore can use TF-IDF to slacken the influence of these speech.
Step 202, data processing equipment 13 utilizes differentiating words identification module 1302, obtains d 1, d 2Differentiating words set.
That is, with this two short text d 1, d 2Compare, obtain the differentiating words set of each short text respectively.Differentiating words be in a short text, occur and in another short text absent variable word.Differentiating words can hint out two relations between the short text usually.
For example, two short texts " price of apple is very high " and " price of banana is very low " are arranged, remove after the stop words, the method for traditional calculating short text similarity can only be found a co-occurrence word " price ".But if can find third party's media, its theme is " fruit ", and so, by " apple " compared with this theme with " banana ", just can find has certain relation between apple and the banana.And then find to have incidence relation between two short texts.
Particularly utilize contrast module 1 in the differentiating words identification module 1302 with this two short text d in this step 1, d 2Compare, and this differentiating words set acquisition module 2 is used for obtaining respectively the differentiating words set of each short text.
The described differentiating words set of obtaining is:
Dist ( d 1 ) = { w | w ∈ d 1 , w ∉ d 2 }
Dist ( d 2 ) = { w | w ∈ d 2 , w ∉ d 1 }
Step 203, select a probability theme, utilize judgment amendment module 1303, judge the differentiating words of this two short text and the correlation degree of probability topic, if the relevance of two differentiating words and same probability topic all is higher than a threshold value, according to the text vector of this this two short text of correlation degree correction.
This probability topic is by extracting among the probability theme collection T, and this probability topic collection is to utilize probability topic abstraction module 1304 to extract by the gibbs sampler method to obtain.Before the step of utilizing the gibbs sampler method to extract probability topic can be executed in step 201, this gibbs sampler was a common practise of the prior art, and related description is seen below continuous the description.Certainly, taking other method of samplings to obtain probability topic also can.
Further comprise following implementation in the step 203:
Step 2031 is utilized theme to choose module 3 and choose a probability topic t from T i
Step 2032 is utilized judge module 4, calculates Dist (d respectively 1), Dist (d 2) in each differentiating words at this probability topic t iOn probability, select short text d 1Differentiating words set at this probability topic t iThe differentiating words w of last probability maximum m, and, short text d 2Differentiating words set at this probability topic t iThe differentiating words w of last probability maximum n
Step 2033 is judged two selected differentiating words w m, w nAt this probability topic t iOn probability whether greater than a preset threshold value λ, if all greater than, represent that these two speech are correlated with, execution in step 2034, otherwise it is uncorrelated to represent them, execution in step 2035.
That is, this step judge to as if p (w m| t i) 〉=λ and p (w n| t i) 〉=λ, p (w m| t i) represent word w mBelong to theme t iProbability.
This preset threshold value λ need obtain by input module 15 inputs.
Step 2034, utilize correcting module 5 according to following formula to V (1), V (2)Revise:
v n ( 1 ) = v n ( 1 ) + v n ( 2 ) × P ( w n | t i )
v m ( 2 ) = v m ( 2 ) + v m ( 1 ) × P ( w m | t i ) ,
Wherein, V n (1)Represent V (1)In n element, V n (2)Represent V (2)In n element, p (w m| t i) represent word w mBelong to theme t iProbability.
That is, as two selected differentiating words w m, w nAt this probability topic t iOn probability during greater than preset threshold value λ, assert this two differentiating words w m, w nHigher with the correlation degree of this probability topic, increase the numerical value of the component in the text vector this moment, to heighten the similarity of two speech, is convenient to the accuracy of two short text similarities of follow-up raising.
Otherwise heighten similarity and also comprise within the scope of the invention, for example, be V (2)In each component increase by a fixed value 0.1, or other numerical value.
Step 2035 utilizes circulation module 6 to judge whether that theme is not selected in addition, if, execution in step 2031, if not, execution in step 204.
Step 204 is utilized computing module 1305, according to the text vector of revised this two short text, utilizes cosine method to calculate short text d 1And d 2Similarity, that is,
Sim ( d 1 , d 2 ) = V ( 1 ) · V ( 2 ) | V ( 1 ) | | V ( 2 ) | ,
Wherein, V (1), V (2)Be d 1And d 2Revised text vector, Sim (d 1, d 2) be d 1And d 2Similarity.
These computational methods only are a kind of example, and the method that other calculate similarity of the prior art also can realize technical scheme of the present invention.
For example, two dot products are calculated similarity, just Sim (d 1, d 2)=V (1)V (2)
Step 205, if still have short text not traveled through in the database, execution in step 201, after all short texts are all traversed, execution in step 206.
Step 206 is utilized sort module 1306, according to this similarity target short text is classified;
In one embodiment, can specifically adopt the k nearest neighbor grader, classify according to this similarity value.The k nearest neighbor grader (KNN, Http:// en.wikipedia.org/wiki/K-nearest neighbor_algorithm) belonging to prior art, the process of its realization also can be referring to above stated specification.
The short text that is all a class is exported.That is, the short text that is all a class is presented on the space of a whole page, perhaps, as the initial value of subsequent treatment.
Above-mentioned steps (201-205) but iterative cycles repeatedly, to confirm in target short text and the database all similarity degrees between the short texts, use and improve the accuracy of classifying.
The extraction of relevant probability topic, the known technology for this area explains slightly at this.In the prior art, utilize the probability topic model (Probabilistic Topic Model, referring to D.M.Blei, A.Y.Ng , ﹠amp; , M.I.Jordan.Latent Dirichlet Allocation.Journal of MachineLearning Research, 3,993-1022,2003.) and represent short text and theme, and word in the text and the relation between the theme.This model is based on following hypothesis, and promptly a short text is made up of a plurality of themes, and each theme is the probability distribution (being that certain word belongs to certain theme) on word.This model can be used for generating short text, and the distribution of an at first selected theme plants at this branch then, to carry out the iteration sampling with each word that generates short text, be shown below,
P ( W ( i ) ) = Σ j = 1 Z P ( W ( i ) | t j ) P ( T ( i ) = t j ) ,
Wherein, P (T (i)=t j) theme t among the document d that generates of expression jProbability; P (W (i)| t j) expression word W (i)Belong to theme t jProbability.
The gibbs sampler method that theme in the probability topic model can propose according to Griffiths and Steyvers is (referring to T.Griffiths and M.Steyvers.Finding scientific topics.The National Academy of Sciences, 101:5228-5235,2004.) extract.This method is at first given each entry a theme at random, then according to following formula to each theme T (i)Carry out the iteration sampling,
P ( T ( i ) = t j | T ( - i ) , W ( i ) , D ( i ) , . ) ∝ C T ( i ) W ( i ) + β Σ W ( i ) C T ( i ) W ( i ) + N · β · C D ( i ) T ( i ) + α Σ T ( i ) C D ( i ) T ( i ) + Z · α
Wherein, P (T (i)=t j| T (-i), W (i), D (i).) and represent assessment to give theme t with entry i jProbability; T (-i)Represent the distribution condition of other all entries; ". " represents other information, as W (-i), D (-i), and super parameter alpha, β.α and β are two parameters that influence subject extraction.
Figure G200910090377XD00113
Represent word W (i)From theme T (i)The number of times of middle sampling does not comprise current entry i;
Figure G200910090377XD00114
Represent theme T (i)Be endowed text D (i)In the number of times of arbitrary entry, do not comprise current entry i.First fraction is represented word W in the formula (i)Belong to theme t jProbability; Second fraction represented theme T (i)At text D (i)The probability that plants of theme branch.
By input module 15 appropriate parameter alpha and β can be set, can obtain one group of probability topic T={t 1..., t i..., t Z, each probability topic all has following form: t i={ t I1, t I2..., t IN, wherein, t IjBe a probability, be used for weighing j speech at t iThe middle probability that occurs.
After subject extraction finishes, word w iBelong to theme t jProbability can be expressed from the next:
P ( w i | t j ) = C t j w i + β Σ i = 1 N C t j w i + N · β .
In another embodiment, the present invention also is used in the question answering system, sends it to the expert place of being familiar with this problem domain automatically at a problem that proposes, and promptly recommends the expert.
Part in question answering system is answered the expert that the user can be set to certain field by the keeper.By calculating the similarity between the problem that this problem short text and expert answered, determine which expert it is recommended.Because the inventive method has improved the accuracy that the short text similarity is calculated, and can give the expert who is familiar with this problem domain with question recommending more accurately.Make question answering system as the platform of information interchange, make the distribution of information more accurate.
See also Fig. 3 for carry out the flow chart of question recommending based on probability topic.
Step 301, a problem short text d of terminal 20 transmission 1Be sent to question answering system server end 10 through communication connection interface 11, the content in data processing equipment 13 these databases 12 of scanning utilizes initialization module 1301 to obtain this short text d 1And obtain a short text d in the short text of all problems of from database 12, storing 2, and to these two short text d 1, d 2Carry out the initialization of text vector respectively.
Its text vector of difference initialization V (1), V (2)For:
V ( 1 ) = { v 1 ( 1 ) , v 2 ( 1 ) , · · · , v N ( 1 ) }
V ( 2 ) = { v 1 ( 2 ) , v 2 ( 2 ) , · · · , v N ( 2 ) } ,
Wherein, v i (j)Be the w in the dictionary iAt short text d jIn weight, this dictionary comprises the word that short text such as all problems and answer comprises in the system; v i (j)Calculate with the TF-IDF method.
Step 302-305 is corresponding with step 202-205, does not give unnecessary details at this.
Step 306 is utilized sort module 1306, according to this similarity Sim (d that calculates 1, d 2) obtain and problem short text d 1K the short text that similarity is the highest obtains and answered this K the expert info of a short text, and it is maximum to judge which expert answers, with problem short text d 1Be sent to this expert's space of a whole page place, finish.
Perhaps, step 306 can also following mode realize, extracts and problem short text d 1The short text d of similarity maximum 2, with short text d 1Be sent to and answered this problem short text d 2Expert's space of a whole page place.
In another embodiment, the present invention also can be used for short text clustering.Such as, in question answering system, to search result clustering, the cluster that perhaps checks on one's answers (referring to the reading and browsing display packing and the system of a Chinese patent application #200510130777.0 problem answers).
Cluster is exactly that semantically close text is gathered to together, as a class.In question answering system, the user submits searching request to one time, generally can return a lot of Search Results.Search Results according to Semantic Clustering, is convenient to the user and is browsed.In like manner, the cluster that also can check on one's answers.
At an assigned short text set (as problem in the question answering system or answer), the cluster flow process is as follows:
(1) utilize each short text in the assigned short text set as target short text successively, and circulation execution in step 201-205, calculate the similarity between any two texts in the assigned short text set;
(2) according to the similarity data that calculate all short texts are carried out cluster.Can adopt in one embodiment K-means clustering algorithm of the prior art ( Http:// en.wikipedia.org/wiki/K-means_clustering) realize.Other clustering methods such as Fuzzy C-means, Hierarchical clustering (http://en.wikipedia.org/wiki/Cluster_Analysis#Fuzzy_c-means_clu stering) etc.
In another embodiment, the present invention is applied in the text search field.
With reference to the structural representation of the search engine server of Fig. 4 A, consult the flow chart that carries out the short text retrieval based on probability topic of Fig. 4 B.
Comprise a communication connection interface 41, database 42, data processing equipment 43, display module 44, input module 45 in the search engine server 40.
Communication connection interface 41 is used to receive the search word from a plurality of terminals 20.Data processing equipment 43 is used for the short text data of search word and database 42 are carried out analyzing and processing.Input module 45 is used for input parameter or supervisory instruction.Display module 44 is used for the short text data after handling are by analysis shown.
In the data processing equipment of search engine server, be provided with the system that carries out the short text classification based on probability topic of the present invention.
Step 401, search subscriber is by terminal input one retrieval short text d 1(being aforementioned search word) to search engine server, the search engine scan database obtains the short text d in the database at random 2, and to these two short text d 1, d 2Carry out the initialization of text vector respectively.
Its text vector of difference initialization V (1), V (2)For:
V ( 1 ) = { v 1 ( 1 ) , v 2 ( 1 ) , · · · , v N ( 1 ) }
V ( 2 ) = { v 1 ( 2 ) , v 2 ( 2 ) , · · · , v N ( 2 ) } ,
Wherein, v i (j)Be the w in the dictionary iAt short text d jIn weight, this dictionary comprises all words in the database short text; v i (j)Calculate with the TF-IDF method.
Step 402-405 is corresponding with step 202-205, does not give unnecessary details at this, during circulation execution in step 401, obtains another short text in the database at random, with retrieval short text d 1Carry out initialized step together; Step 406, after all short texts in the database are all traversed, extract similarity greater than the short text in the database 42 of predetermined value as result for retrieval, just sorting according to similarity is sent to user terminal to show.
Search method of the present invention is based on probability topic and retrieves, and whether reaches threshold value to judge its correlation in terms of content by judging similarity, has improved the accuracy of short text search return results greatly.
In another embodiment, the present invention is applied in mail server to filter spam.
Be provided with the system that carries out the short text classification based on probability topic of the present invention in the data processing equipment of mail server.
With reference to Fig. 5 A is the structural representation of mail server, consults Fig. 5 B for discern the flow chart of spam based on probability topic.
Comprise a communication connection interface 51, email storage database 52, data processing equipment 53, display module 54, input module 55 in the mail server 50.
Communication connection interface 51 and a plurality of terminal 20 transceive data.The mail of newly receiving is stored in the email storage database 52.Data processing equipment 53 is used for the mail short text of email storage database 52 is carried out analyzing and processing.Input module 55 is used for input parameter or supervisory instruction.Display module 54 is used for processed mail short text data are by analysis shown.
In the email storage database 52, comprise the model memory module that stores some spam title short text models, and store the email storage module that newly receives mail.
Step 501, the content in the email storage database of scan mail server is obtained a model short text d from the model memory module 1, from the email storage module, obtain a mail title short text d more at random or in proper order 2, and to these two short text d 1, d 2Carry out the initialization of text vector respectively.
Its text vector of difference initialization V (1), V (2)For:
V ( 1 ) = { v 1 ( 1 ) , v 2 ( 1 ) , · · · , v N ( 1 ) }
V ( 2 ) = { v 1 ( 2 ) , v 2 ( 2 ) , · · · , v N ( 2 ) } ,
Wherein, v i (j)Be the w in the dictionary iAt short text d jIn weight, this dictionary comprises all words in the email storage database mail title; v i (j)Calculate with the TF-IDF method.
Step 502-505 is corresponding with step 202-205.
When step 505 circulation execution in step 501, from the email storage module, obtain another mail title short text and model short text d at random or in proper order 1Execution in step, the short text in the email storage module is all traveled through, and obtains next model short text again and repeat abovementioned steps from the model memory module.
Step 506, after all short texts and model short text are all traversed, extract with the model memory module in the similarity of short text greater than the mail title short text of predetermined value, shield or do special sign, be spam to show it.Or according to the spam that identifies, the operation that shields.
Should special sign comprise concentrated the demonstration, be designated as highlighted state, be designated as particular color.
Spam recognition methods of the present invention is based on probability topic to be carried out, by judging that similarity to judge its correlation in terms of content, has therefore also improved the accuracy of return results greatly.
The present invention also can be used for the rubbish of portable terminal, the shielding of advertising message.
As shown in Figure 6A for the structural representation of the present invention's one Application Example.
One mobile communication system comprises a portable terminal 20, it for example is a mobile phone, portable terminal 20 is connected with a short message server 60 by a fire compartment wall 61, the system that carries out short text classification based on probability topic of the present invention is arranged in this fire compartment wall, and the note that is used for short message server 60 is sent to portable terminal 20 is tackled, screened.Particularly the keeper is kept at model memory module in this fire compartment wall with the typical model of the note of desire interception.
Be depicted as the schematic flow sheet of the present invention's one Application Example as Fig. 6 B.
Step 601, fire compartment wall 61 receives the current short message text d that sends from short message server 60 1, and short messages stored model in the scanning model memory module, and from the model memory module, obtain a model short text d 2, and to these two short text d 1, d 2Carry out the initialization of text vector respectively.
Its text vector of difference initialization V (1), V (2)For:
V ( 1 ) = { v 1 ( 1 ) , v 2 ( 1 ) , · · · , v N ( 1 ) }
V ( 2 ) = { v 1 ( 2 ) , v 2 ( 2 ) , · · · , v N ( 2 ) } ,
Wherein, v i (j)Be the w in the dictionary iAt short text d jIn weight, this dictionary comprises all words in note set of prior structure, these notes comprise normal note and refuse messages; v i (j)Calculate with the TF-IDF method.
Step 602-605 is corresponding with step 202-205.
In case find current short message text d 1Be higher than a predetermined value with the similarity of arbitrary model short text, tackle this short message text d 1, do not mail to portable terminal 20.
If do not find current short message text d yet 1Be higher than a predetermined value with the similarity of whole model short texts, current short message text is sent to portable terminal 20.
By said method can be tight interception identical in form with the model short text or on meaning relevant note, make that the examination of refuse messages is tighter.
Perhaps, fire compartment wall 61 also can be connected with a Internet web server 62, and the user of this portable terminal 20 can login this server in station 62 and bind the number of this portable terminal 20.The user can pass through this server in station 62, and the model short text in this fire compartment wall 61 is set.That is the short message type that these portable terminal 20 needs interceptions are set of personalization.Make the interception and the clearance of note, satisfy user's demand more.
In one embodiment of the invention, the inventive method and system can be applied to the problem classification in the question answering system.Problem is a kind of special short text, existing research work mostly at the classification of type of problem (as X.Li, D.Roth:Learning Question Classifiers.In:Proceedings ofthe 19th International Conference on Computational Linguistics (2002), D.Zhang, W.S.Lee:Question Classification using Support VectorMachine.In:Proceedings of the 26th annual international ACM SIGIRconference on Research and development in informaion retrieval (2003)), just classify according to the potential answer type of problem, such as the personage, place etc.But a lot of on-line systems (as based on interactively question answering system) need be organized the problem of magnanimity according to the theme of problem, promptly according to the content of problem it are assigned to corresponding classification, as computer, and education, motion etc.In order to realize that problem is classified, many graders (k nearest neighbor grader for example, KNN) need similarity between the computational problem, and the short text calculation of similarity degree method based on probability topic that the present invention proposes can come with the machine learning method of this classics problem is classified.The inventive method also can be verified by classification performance.
In one embodiment of the invention, having collected two data sets tests.A data set comprises 1120 problems in 32 classifications from BuyAns, and the number of problem from 14 to 108 does not wait in each classification; Answers comprises 2400 problems in 11 classifications, and it is minimum 100 that each classification comprises, maximum 400 problems.Behind the stop words in removing problem, all words are converted to prototype (at English word, as removing forms such as past tense).The English word base that each problem comprises originally is less than 10.In each data centralization, 70% the problem randomly drawed is as training set, and remaining is tested.Average result is write down in retest 10 times at last.Because the gibbs sampler algorithm can both obtain target distribution later 300 times in circulation usually, so cycle-index is made as 300.α and β rule of thumb are made as 50/Z and 0.01 respectively.
In one embodiment of the invention, the performance of following three kinds of graders: (1) KNN; (2) SVM; (3) KNN combines (KNN_TBS) with the short text similarity calculating method based on probability topic of the present invention.Therefore KNN and SVM are based on the TF-IDF vector representation, as the comparison of the inventive method.In KNN and KNN_TBS, neighbours' quantity is made as 30.SVM is based on libSVM instrument (C.Chang andC.Lin, LIBSVM:a library for support vector machines, 2001.Softwareavailable at:http: //www.csie.ntu.edu.tw/cjlin/libsvm/), in experiment, adopted the linear kernel function of SVM.Basic skills such as accuracy, recall rate and F1 are used to the above three kinds of performances that grader is classified to problem of evaluate application.Calculated the little average and grand average of accuracy, recall rate and F1 when as follows.
In the methods of the invention, the number of topics of extraction is an important parameter, can optimize according to classification performance.Shown in Fig. 7 A, 7B, be on two data sets, get little average (MicroF1) and the grand mean value (MacroF1) of the F1 that the different themes number obtains. On two data sets of Answer, when number of topics was respectively 160 and 120, classification performance was the highest.
Equally, λ also is an important parameter, and it determines whether distinguish word set for two is correlated with by probability topic.If select too muchly, might neglect the true association between the branch word set; Select too smallly, then can misjudge the association of distinguishing between the word set.Shown in Fig. 8 A, 8B, be on two data sets, get little average (MicroF1) and the grand mean value (MacroF1) of the F1 that different λ values obtain.
Shown in table 1, table 2, be respectively the test result of two data sets.Wherein, the number of topics of the extraction of BuyAns data set is made as 160, and the λ value is made as 0.05; The number of topics of the extraction of Answer data set is made as 120, and the λ value is made as 0.15.As can be seen from the table, in conjunction with method of the present invention, compare with traditional TF-IDF method, classify accuracy has improved 10% on the BuyAns data set, compare with SVM, and majority also increases. On the Answer data set, compare with traditional TF-IDF method, the little average and grand average mark of F1 you can well imagine high 12% and 17%, compares with SVM, has also improved 10%.
The problem class test result of table 1 BuyAns data set
Figure G200910090377XD00171
The problem class test result of Answer data set
Figure G200910090377XD00181
The foregoing description only for explanation the present invention's usefulness, is not considered as limitation of the present invention, and concrete protection range please be looked the appended claim book and be as the criterion.

Claims (17)

1, a kind of method of carrying out the short text classification based on probability topic is applied in the data processing equipment of question answering system, is used for according to the similarity between short text short text being classified, and it is characterized in that, comprises the steps:
According to a target short text and a short text that is obtained from the database of question answering system of input, obtain its initialized text vector respectively;
Scan this two short text, obtain the differentiating words of this two short text respectively;
When the correlation degree of the differentiating words of this two short text and a probability theme all is higher than threshold value, according to the text vector of this this two short text of correlation degree correction;
Calculate similarity between two short texts according to the text vector of revised this two short text;
Obtain another short text in the database of question answering system, the short text in database is finished by traversal, carries out this scanning step;
According to this similarity target short text is classified.
2, the method for claim 1 is characterized in that, obtains initialized text vector by scanning this two short text: V ( 1 ) = { v 1 ( 1 ) , v 2 ( 1 ) , . . . , v N ( 1 ) } V ( 2 ) = { v 1 ( 2 ) , v 2 ( 2 ) , . . . , v N ( 2 ) } ,
Wherein, v (j)Be text d jInitialized text vector, v i (j)Be i speech w in the dictionary iAt short text d jIn weight, this dictionary comprises all words in all short texts in the question answering system.
3, method as claimed in claim 1 or 2 is characterized in that, describedly further comprises by scanning the step that this two short text obtains the differentiating words of this two short text respectively:
This two short text is compared, obtain the differentiating words set of each short text respectively, described differentiating words set is:
Dist ( d 1 ) = { w | w ∈ d 1 , w ∉ d 2 }
Dist ( d 2 ) = { w | w ∈ d 2 , w ∉ d 1 }
Wherein, d iBe short text, Dist (d i) be short text d iDifferentiating words set, w is a word.
4, the method for claim 1 is characterized in that, also comprises the step of utilizing the gibbs sampler method to extract this probability topic.
5, method as claimed in claim 3 is characterized in that, the described differentiating words and the correlation degree of a probability theme when all being higher than threshold value when this two short text further comprises according to the step of the text vector of this this two short text of correlation degree correction:
Step 31 is chosen a probability topic t i, search for Dist (d respectively 1), Dist (d 2) at this probability topic t iThe differentiating words of last probability maximum;
Step 32 judges that selected two differentiating words are at probability topic t iOn probability whether all greater than threshold value, if all greater than, execution in step 33; Otherwise, execution in step 34;
Step 33, according to following formula to V (1), V (2)Revise:
v n ( 1 ) = v n ( 1 ) + v n ( 2 ) × P ( w n | t i ) v m ( 2 ) = v m ( 2 ) + v m ( 1 ) × P ( w m | t i ) , Wherein, V n (1)Represent V (1)In n element, V n (2)Represent V (2)In n element, p (w m| t i) represent word w mBelong to theme t iProbability, word w mBe short text d 1In at probability topic t iThe word of last probability maximum, w nBe short text d 2In at probability topic t iThe word of last probability maximum.
Step 34 judges whether that theme is not selected in addition, if execution in step 31 if not, is carried out the step that described text vector according to revised this two short text calculates the similarity between two short texts.
6, method as claimed in claim 5 is characterized in that, the step that described text vector according to this two short text calculates the similarity between two short texts comprises:
Sim ( d 1 , d 2 ) = V ( 1 ) · V ( 2 ) | V ( 1 ) | | V ( 2 ) | ,
Wherein, Sim (d 1, d 2) be d 1And d 2Similarity, V (1), V (2)Be d 1And d 2Revised text vector.
7, the method for claim 1 is characterized in that, the step of described classification utilizes k nearest neighbor grader or centroid method to realize.
8, a kind of system that carries out the short text classification based on probability topic is applied in the data processing equipment of question answering system, is used for according to the similarity between short text short text being classified, and it is characterized in that, comprising:
Initialization module, a target short text and a short text that is obtained from the database of question answering system according to input obtain its initialized text vector respectively;
The differentiating words identification module is used to scan this two short text, and obtains the differentiating words of this two short text respectively;
Judgment amendment module is when the correlation degree of the differentiating words of this two short text and a probability theme all is higher than threshold value, according to the text vector of this this two short text of correlation degree correction;
Computing module calculates similarity between two short texts according to the text vector of revised this two short text;
Sort module is classified to target short text according to similarity.
9, system as claimed in claim 8 is characterized in that, described initialization module by the initialized text vector that scans this two short texts acquisition is: V ( 1 ) = { v 1 ( 1 ) , v 2 ( 1 ) , . . . , v N ( 1 ) } V ( 2 ) = { v 1 ( 2 ) , v 2 ( 2 ) , . . . , v N ( 2 ) } ,
Wherein, v (j)Be text d jInitialized text vector, w i (j)Be i speech w in the dictionary iAt short text d jIn weight, this dictionary comprises all words in all short texts in the question answering system.
10, system as claimed in claim 8 or 9 is characterized in that described differentiating words identification module further comprises: contrast module and differentiating words set acquisition module;
This contrast module is used for this two short text is compared;
This differentiating words set acquisition module is used for obtaining respectively the differentiating words set of each short text, and described differentiating words set is:
Dist ( d 1 ) = { w | w ∈ d 1 , w ∉ d 2 }
Dist ( d 2 ) = { w | w ∈ d 2 , w ∉ d 1 }
Wherein, d iBe short text, Dist (d i) be short text d iDifferentiating words set, w is a word.
11, system as claimed in claim 8 is characterized in that, also comprises a probability subject extraction module, and this probability topic abstraction module utilizes the gibbs sampler method to extract this probability topic.
12, system as claimed in claim 10 is characterized in that, described judgment amendment module further comprises:
Module chosen in theme, is used to choose a probability topic t i, search for Dist (d respectively 1), Dist (d 2) at this probability topic t iThe speech of last probability maximum;
Judge module is used to judge that selected two differentiating words are at probability topic t iOn probability whether all greater than threshold value, if all greater than, carry out correcting module; Otherwise, carry out the circulation module;
Correcting module is used for according to following formula V (1), V (2)Revise:
v n ( 1 ) = v n ( 1 ) + v n ( 2 ) × P ( w n | t i ) v m ( 2 ) = v m ( 2 ) + v m ( 1 ) × P ( w m | t i ) , Wherein, V n (1)Represent V (1)In n element, V n (2)Represent V (2)In n element, p (w m| t i) represent word w mBelong to theme t iProbability, word w mBe short text d 1In at probability topic t iThe word of last probability maximum, w nBe short text d 2In at probability topic t iThe word of last probability maximum.
The circulation module is used to judge whether also have theme not to be selected, if, carry out theme and choose module, if not, carry out computing module.
13, method as claimed in claim 12 is characterized in that, described computing module utilizes following mode to realize:
Sim ( d 1 , d 2 ) = V ( 1 ) · V ( 2 ) | V ( 1 ) | | V ( 2 ) | ,
Wherein, Sim (d 1, d 2) be d 1And d 2Similarity, V (1), V (2)Be d 1And d 2Revised text vector.
14, method as claimed in claim 8 is characterized in that, described sort module utilizes k nearest neighbor grader or centroid method to realize.
15, a kind of method of carrying out the short text retrieval based on probability topic is applied to it is characterized in that in the search engine server, comprises the steps:
Step 1 receives the retrieval short text as term that transmits from the search subscriber terminal;
Step 2 is obtained a short text from the database of search engine server;
Step 3 according to above-mentioned retrieval short text and the short text that obtains, is obtained its initialized text vector respectively from database;
Step 4 scans this two short text, obtains the differentiating words of this two short text respectively;
Step 5 is when the correlation degree of the differentiating words of this two short text and a probability theme all is higher than threshold value, according to the text vector of this this two short text of correlation degree correction;
Step 6 is calculated similarity between two short texts according to the text vector of revised this two short text;
Step 7, circulation execution in step 2, the short text traversal in database finishes;
Step 8 is extracted and the short text of the similarity of retrieving short text greater than predetermined value, is sent to the search subscriber terminal and shows.
16, a kind of method of the identification spam based on probability topic is applied to it is characterized in that in the mail server, comprises the steps:
Step 1, storage model spam short text in the database of mail server, and obtain one of them;
Step 2 is obtained a short text from the database of mail server;
Step 3 according to above-mentioned two short texts, is obtained its initialized text vector respectively;
Step 4 scans this two short text, obtains the differentiating words of this two short text respectively;
Step 5 is when the correlation degree of the differentiating words of this two short text and a probability theme all is higher than threshold value, according to the text vector of this this two short text of correlation degree correction;
Step 6 is calculated similarity between two short texts according to the text vector of revised this two short text;
Step 7, the circulation execution in step 2 whole short text traversals in database finish, and after the whole short text traversals in database finished, execution in step 1 all traveled through until model spam short text and finishes again;
Step 8 will be designated spam greater than the short text that obtains of predetermined value with the similarity of this model spam short text from the database of mail server.
17, a kind of method based on shielding rubbish, advertising message in the portable terminal of probability topic, be applied in the fire compartment wall of mobile communication system, store the model short text in the fire compartment wall, mobile communication system comprises short message server, fire compartment wall, user terminal, it is characterized in that, comprise the steps:
Step 1, fire compartment wall receive the short message text that this short message server sends;
Step 2, fire compartment wall are obtained a model short text;
Step 3 according to this short message text and this model short text, is obtained its initialized text vector respectively;
Step 4 scans this two short text, obtains the differentiating words of this two short text respectively;
Step 5 is when the correlation degree of the differentiating words of this two short text and a probability theme all is higher than threshold value, according to the text vector of this this two short text of correlation degree correction;
Step 6 is calculated similarity between two short texts according to the text vector of revised this two short text;
Step 7, circulation execution in step 2, all the model short text traversals in fire compartment wall finish;
Step 8, the similarity of fire compartment wall interception and model short text is higher than this short message text of a predetermined value.
CN200910090377A 2009-08-07 2009-08-07 Method and system for classifying short texts based on probability topic Pending CN101621391A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN200910090377A CN101621391A (en) 2009-08-07 2009-08-07 Method and system for classifying short texts based on probability topic

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN200910090377A CN101621391A (en) 2009-08-07 2009-08-07 Method and system for classifying short texts based on probability topic

Publications (1)

Publication Number Publication Date
CN101621391A true CN101621391A (en) 2010-01-06

Family

ID=41514455

Family Applications (1)

Application Number Title Priority Date Filing Date
CN200910090377A Pending CN101621391A (en) 2009-08-07 2009-08-07 Method and system for classifying short texts based on probability topic

Country Status (1)

Country Link
CN (1) CN101621391A (en)

Cited By (26)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN102073704A (en) * 2010-12-24 2011-05-25 华为终端有限公司 Text classification processing method, system and equipment
CN102681983A (en) * 2011-03-07 2012-09-19 北京百度网讯科技有限公司 Alignment method and device for text data
CN103049433A (en) * 2012-12-11 2013-04-17 微梦创科网络科技(中国)有限公司 Automatic question answering method, automatic question answering system and method for constructing question answering case base
CN103457975A (en) * 2012-06-01 2013-12-18 腾讯科技(深圳)有限公司 Method and device for acquiring map interest point evaluation data
CN103823809A (en) * 2012-11-16 2014-05-28 百度在线网络技术(北京)有限公司 Query phrase classification method and device, and classification optimization method and device
CN103874033A (en) * 2012-12-12 2014-06-18 上海粱江通信系统股份有限公司 Method for identifying irregular spam short message on the basis of Chinese word segmentation
CN104461224A (en) * 2013-09-23 2015-03-25 联想(北京)有限公司 Information processing method and electronic device
CN105005564A (en) * 2014-04-17 2015-10-28 北京搜狗科技发展有限公司 Data processing method and apparatus based on question-and-answer platform
WO2015165230A1 (en) * 2014-04-28 2015-11-05 华为技术有限公司 Social contact message monitoring method and device
CN106095845A (en) * 2016-06-02 2016-11-09 腾讯科技(深圳)有限公司 File classification method and device
CN106156142A (en) * 2015-04-13 2016-11-23 深圳市腾讯计算机系统有限公司 The processing method of a kind of text cluster, server and system
CN106649255A (en) * 2015-11-04 2017-05-10 江苏引跑网络科技有限公司 Method for automatically classifying and identifying subject terms of short texts
CN106782516A (en) * 2016-11-17 2017-05-31 北京云知声信息技术有限公司 Language material sorting technique and device
CN107315731A (en) * 2016-04-27 2017-11-03 北京京东尚科信息技术有限公司 Text similarity computing method
CN107533574A (en) * 2015-09-03 2018-01-02 华为技术有限公司 Email relationship finger system based on random index pattern match
CN108345424A (en) * 2018-01-31 2018-07-31 维沃移动通信有限公司 A kind of method for information display and mobile terminal
CN109063032A (en) * 2018-07-16 2018-12-21 清华大学 A kind of noise-reduction method of remote supervisory retrieval data
CN109783127A (en) * 2018-11-26 2019-05-21 中国船舶重工集团公司第七0九研究所 A kind of the demand allocation problem method for solving and device of serviceization information system
CN110019794A (en) * 2017-11-07 2019-07-16 腾讯科技(北京)有限公司 Classification method, device, storage medium and the electronic device of textual resources
CN110069772A (en) * 2019-03-12 2019-07-30 平安科技(深圳)有限公司 Predict device, method and the storage medium of the scoring of question and answer content
CN110413753A (en) * 2019-07-22 2019-11-05 阿里巴巴集团控股有限公司 The extended method and device of question and answer sample
CN110619117A (en) * 2018-06-19 2019-12-27 广州优视网络科技有限公司 Keyword extraction method and device
CN111125334A (en) * 2019-12-20 2020-05-08 神思电子技术股份有限公司 Search question-answering system based on pre-training
WO2020132933A1 (en) * 2018-12-25 2020-07-02 深圳市优必选科技有限公司 Short text filtering method and apparatus, medium and computer device
CN112988954A (en) * 2021-05-17 2021-06-18 腾讯科技(深圳)有限公司 Text classification method and device, electronic equipment and computer-readable storage medium
US11100412B2 (en) 2019-07-22 2021-08-24 Advanced New Technologies Co., Ltd. Extending question and answer samples

Cited By (43)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN102073704A (en) * 2010-12-24 2011-05-25 华为终端有限公司 Text classification processing method, system and equipment
CN102073704B (en) * 2010-12-24 2013-09-25 华为终端有限公司 Text classification processing method, system and equipment
CN102681983A (en) * 2011-03-07 2012-09-19 北京百度网讯科技有限公司 Alignment method and device for text data
CN103457975A (en) * 2012-06-01 2013-12-18 腾讯科技(深圳)有限公司 Method and device for acquiring map interest point evaluation data
CN103457975B (en) * 2012-06-01 2016-08-31 腾讯科技(深圳)有限公司 The method and apparatus obtaining map interest point evaluation data
CN103823809A (en) * 2012-11-16 2014-05-28 百度在线网络技术(北京)有限公司 Query phrase classification method and device, and classification optimization method and device
CN103823809B (en) * 2012-11-16 2018-06-08 百度在线网络技术(北京)有限公司 A kind of method, the method for Classified optimization and its device to query phrase classification
CN103049433A (en) * 2012-12-11 2013-04-17 微梦创科网络科技(中国)有限公司 Automatic question answering method, automatic question answering system and method for constructing question answering case base
CN103049433B (en) * 2012-12-11 2015-10-28 微梦创科网络科技(中国)有限公司 The method of automatic question-answering method, automatically request-answering system and structure question and answer case library
CN103874033A (en) * 2012-12-12 2014-06-18 上海粱江通信系统股份有限公司 Method for identifying irregular spam short message on the basis of Chinese word segmentation
CN103874033B (en) * 2012-12-12 2017-11-24 上海粱江通信系统股份有限公司 A kind of method that irregular refuse messages are identified based on Chinese word segmentation
CN104461224A (en) * 2013-09-23 2015-03-25 联想(北京)有限公司 Information processing method and electronic device
CN105005564A (en) * 2014-04-17 2015-10-28 北京搜狗科技发展有限公司 Data processing method and apparatus based on question-and-answer platform
CN105005564B (en) * 2014-04-17 2019-09-03 北京搜狗科技发展有限公司 A kind of data processing method and device based on answer platform
US10250550B2 (en) 2014-04-28 2019-04-02 Huawei Technologies Co., Ltd. Social message monitoring method and apparatus
WO2015165230A1 (en) * 2014-04-28 2015-11-05 华为技术有限公司 Social contact message monitoring method and device
CN106156142A (en) * 2015-04-13 2016-11-23 深圳市腾讯计算机系统有限公司 The processing method of a kind of text cluster, server and system
CN106156142B (en) * 2015-04-13 2020-06-02 深圳市腾讯计算机系统有限公司 Text clustering processing method, server and system
CN107533574A (en) * 2015-09-03 2018-01-02 华为技术有限公司 Email relationship finger system based on random index pattern match
US10936638B2 (en) 2015-09-03 2021-03-02 Huawei Technologies Co., Ltd. Random index pattern matching based email relations finder system
CN106649255A (en) * 2015-11-04 2017-05-10 江苏引跑网络科技有限公司 Method for automatically classifying and identifying subject terms of short texts
CN107315731A (en) * 2016-04-27 2017-11-03 北京京东尚科信息技术有限公司 Text similarity computing method
CN106095845A (en) * 2016-06-02 2016-11-09 腾讯科技(深圳)有限公司 File classification method and device
CN106782516B (en) * 2016-11-17 2020-02-07 北京云知声信息技术有限公司 Corpus classification method and apparatus
CN106782516A (en) * 2016-11-17 2017-05-31 北京云知声信息技术有限公司 Language material sorting technique and device
CN110019794B (en) * 2017-11-07 2023-04-25 腾讯科技(北京)有限公司 Text resource classification method and device, storage medium and electronic device
CN110019794A (en) * 2017-11-07 2019-07-16 腾讯科技(北京)有限公司 Classification method, device, storage medium and the electronic device of textual resources
CN108345424A (en) * 2018-01-31 2018-07-31 维沃移动通信有限公司 A kind of method for information display and mobile terminal
CN110619117B (en) * 2018-06-19 2024-03-19 阿里巴巴(中国)有限公司 Keyword extraction method and device
CN110619117A (en) * 2018-06-19 2019-12-27 广州优视网络科技有限公司 Keyword extraction method and device
CN109063032B (en) * 2018-07-16 2020-09-11 清华大学 Noise reduction method for remote supervision and retrieval data
CN109063032A (en) * 2018-07-16 2018-12-21 清华大学 A kind of noise-reduction method of remote supervisory retrieval data
CN109783127B (en) * 2018-11-26 2022-05-24 中国船舶重工集团公司第七0九研究所 Demand configuration problem solving method and device for service information system
CN109783127A (en) * 2018-11-26 2019-05-21 中国船舶重工集团公司第七0九研究所 A kind of the demand allocation problem method for solving and device of serviceization information system
WO2020132933A1 (en) * 2018-12-25 2020-07-02 深圳市优必选科技有限公司 Short text filtering method and apparatus, medium and computer device
CN110069772B (en) * 2019-03-12 2023-10-20 平安科技(深圳)有限公司 Device, method and storage medium for predicting scoring of question-answer content
CN110069772A (en) * 2019-03-12 2019-07-30 平安科技(深圳)有限公司 Predict device, method and the storage medium of the scoring of question and answer content
CN110413753B (en) * 2019-07-22 2020-09-22 阿里巴巴集团控股有限公司 Question-answer sample expansion method and device
US11100412B2 (en) 2019-07-22 2021-08-24 Advanced New Technologies Co., Ltd. Extending question and answer samples
CN110413753A (en) * 2019-07-22 2019-11-05 阿里巴巴集团控股有限公司 The extended method and device of question and answer sample
CN111125334B (en) * 2019-12-20 2023-09-12 神思电子技术股份有限公司 Search question-answering system based on pre-training
CN111125334A (en) * 2019-12-20 2020-05-08 神思电子技术股份有限公司 Search question-answering system based on pre-training
CN112988954A (en) * 2021-05-17 2021-06-18 腾讯科技(深圳)有限公司 Text classification method and device, electronic equipment and computer-readable storage medium

Similar Documents

Publication Publication Date Title
CN101621391A (en) Method and system for classifying short texts based on probability topic
Mallick et al. Digital media news categorization using Bernoulli document model for web content convergence
CN109885773B (en) Personalized article recommendation method, system, medium and equipment
Ghanbari-Adivi et al. Text emotion detection in social networks using a novel ensemble classifier based on Parzen Tree Estimator (TPE)
US7689531B1 (en) Automatic charset detection using support vector machines with charset grouping
KR102069621B1 (en) Apparatus and Method for Documents Classification Using Documents Organization and Deep Learning
US7711673B1 (en) Automatic charset detection using SIM algorithm with charset grouping
Lee et al. Engineering doc2vec for automatic classification of product descriptions on O2O applications
Gupta et al. A novel hybrid text summarization system for Punjabi text
CN111797239A (en) Application program classification method and device and terminal equipment
CN104361037A (en) Microblog classifying method and device
CN101211368B (en) Method for classifying search term, device and search engine system
Fu et al. Open-categorical text classification based on multi-LDA models
Shrivastava et al. Enhancing aggression detection using GPT-2 based data balancing technique
Puteh et al. Sentiment mining of Malay newspaper (SAMNews) using artificial immune system
Escalante et al. Particle swarm model selection for authorship verification
Gupta et al. Fake news detection using machine learning
Al Mostakim et al. Bangla content categorization using text based supervised learning methods
Mani et al. Email spam detection using gated recurrent neural network
Illig et al. A comparison of content-based tag recommendations in folksonomy systems
CN112487263A (en) Information processing method, system, equipment and computer readable storage medium
Moumtzidou et al. Discovery of environmental nodes in the web
Rahman et al. An efficient deep learning technique for bangla fake news detection
CN115510269A (en) Video recommendation method, device, equipment and storage medium
Reddy et al. Classification of Spam Text using SVM

Legal Events

Date Code Title Description
C06 Publication
PB01 Publication
C10 Entry into substantive examination
SE01 Entry into force of request for substantive examination
C02 Deemed withdrawal of patent application after publication (patent law 2001)
WD01 Invention patent application deemed withdrawn after publication

Open date: 20100106