CN104915359A - Theme label recommending method and device - Google Patents

Theme label recommending method and device Download PDF

Info

Publication number
CN104915359A
CN104915359A CN201410096024.1A CN201410096024A CN104915359A CN 104915359 A CN104915359 A CN 104915359A CN 201410096024 A CN201410096024 A CN 201410096024A CN 104915359 A CN104915359 A CN 104915359A
Authority
CN
China
Prior art keywords
theme
message
vocabulary
label
user
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN201410096024.1A
Other languages
Chinese (zh)
Other versions
CN104915359B (en
Inventor
佘洁莹
陈雷
梁颖琪
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Honor Device Co Ltd
Original Assignee
Huawei Technologies Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Huawei Technologies Co Ltd filed Critical Huawei Technologies Co Ltd
Priority to CN201410096024.1A priority Critical patent/CN104915359B/en
Publication of CN104915359A publication Critical patent/CN104915359A/en
Application granted granted Critical
Publication of CN104915359B publication Critical patent/CN104915359B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Landscapes

  • Information Retrieval, Db Structures And Fs Structures Therefor (AREA)

Abstract

The embodiment of the invention provides a theme label recommending method and device. The theme label recommending method comprises the steps that a first message published by a user on a social network is obtained, wherein the first message is a message which is not labeled with a theme label; according to a theme model of the social network, the theme labels of the first message are obtained; the obtained theme labels of the first message are displayed on a screen to be selected by a user; according to the selection result of the user, the theme label of the first message is determined. By the adoption of the theme label recommending method and the device, the purpose that theme labels are recommended to the user according to the theme model, the theme label is determined finally according to selection of the user, and therefore the finally determined theme label is closest to the theme which the user wants to express.

Description

Theme label recommend method and device
Technical field
The present invention relates to the information processing technology, particularly relate to a kind of theme label recommend method and device.
Background technology
Universal along with Internet and mobile terminal, uses the user of social networks to get more and more.Everybody can issue or forward some message by social networks, and along with increasing of user, the size of message that social networks is issued is also increasing.
In order to promote the experience of user on social networks, existing social networks proposes theme label (hashtag) function, namely isolate vocabulary to demarcate segregate vocabulary as hashtag by " # " number when user gives out information, this hashtag is the theme that given out information.Such hashtag can help user to obtain the message such as specific topics, dependent event.But this mode must rely on when user gives out information and all incidentally go up assigned tags to demarcate hashtag at every turn, if user does not demarcate hashtag just can not determine out the theme that user gives out information.
For the message of not demarcating hashtag, determine the mode of message subject in prior art, main employing message cluster, is occurred that a fairly large number of vocabulary is as theme label in the message that user is issued.
But the theme label adopting this clustering technique to obtain often accurately can not express the theme that user gives out information.
Summary of the invention
The embodiment of the present invention provides a kind of theme label recommend method and device, and the theme label for solving recommendation accurately can not express the problem of the theme that user gives out information.
Embodiment of the present invention first aspect provides a kind of theme label recommend method, comprising:
Obtain the first message that user issues on social networks, described first message refers to the message not being labeled theme label;
According to the topic model of described social networks, obtain the theme label of described first message;
The theme label of the first message of described acquisition is presented on screen and selects for described user;
According to the result that described user selects, determine the theme label of described first message.
In conjunction with first aspect, in the first possible embodiment of first aspect, the described topic model according to described social networks, before obtaining the theme label of described first message, also comprises:
From described social networks, obtain the second message, described second message refers to the message being marked with theme label;
Pre-service is carried out to described second message, and preserves described pretreated result;
By using machine learning method to train described pretreated result, obtain the topic model of described social networks.
In conjunction with the first possible embodiment of first aspect, in the embodiment that the second of first aspect is possible, described pre-service is carried out to described second message, and preserves described pretreated result, comprising:
Described second message is carried out cutting by vocabulary;
From the result of described cutting, phrase is obtained according to default part of speech;
Described phrase is stored according to preset format.
In conjunction with the first possible embodiment of first aspect or the possible embodiment of the second, in the embodiment that the second of first aspect is possible, described by using machine learning method to train described pretreated result, the topic model obtaining described social networks comprises:
Use machine learning method to sample to described pretreated result, obtain theme-vocabulary probability distribution and theme-theme label probability distribution according to described sampled result;
According to described theme-vocabulary probability distribution and described theme-theme label probability distribution, obtain the topic model of described social networks.
In conjunction with first aspect to first aspect the third possible embodiment any one of, in the 4th kind of possible embodiment of first aspect, the described topic model according to described social networks, the theme label obtaining described first message comprises:
Adopt formula p ( h | d u ) = Σ t ψ t , h × ( ( Σ r ∈ F u ∪ { U } η u , r θ r , t ) Π w u , d , n ( π 0 φ B , w u , d , n + π 1 φ t , w u , d , n ) ) Calculate described first message d utheme label be the probability of h, and the value obtaining described p (hdu) is greater than the theme label of the first predetermined threshold value, wherein: ψ t,hrepresent the probability occurring theme label h in the theme label of theme t, η u,rrepresent that described user selects the probability of the history theme of user r, θ r,trepresent that the message of user r issue belongs to the probability of theme t, π 0represent that a vocabulary belongs to the probability of not a theme vocabulary, represent in not a theme vocabulary and occur w u, d, nprobability, π 1represent that a vocabulary belongs to the probability of theme vocabulary, represent to belong in the vocabulary of theme t and occur w u, d, nprobability, w u, d, nrepresent the vocabulary n in described first message;
The described theme label by the first message of described acquisition is presented on screen and selects to comprise for described user:
By described p (h|d u) the value theme label that is greater than the first predetermined threshold value be presented on screen and select for described user.
In conjunction with the 4th kind of possible embodiment of first aspect, in the 5th kind of possible embodiment of first aspect, described employing formula p ( h | d u ) = Σ t ψ t , h × ( ( Σ r ∈ F u ∪ { U } η u , r θ r , t ) Π w u , d , n ( π 0 φ B , w u , d , n + π 1 φ t , w u , d , n ) ) Calculate described first message d utheme label be the probability of h, comprising:
By the ψ in the topic model of described social networks t,harrange from big to small, and according to ψ t,horder from big to small substitutes into formula successively p ( h | d u ) = Σ t ψ t , h × ( ( Σ r ∈ F u ∪ { U } η u , r θ r , t ) Π w u , d , n ( π 0 φ B , w u , d , n + π 1 φ t , w u , d , n ) ) Calculate, when calculating the p (h|d obtained u) value stops calculating when being less than the second predetermined threshold value, described second predetermined threshold value is less than or equal to described first predetermined threshold value.
Embodiment of the present invention second aspect provides a kind of theme label recommendation apparatus, comprising:
Acquisition module, for obtaining the first message that user issues on social networks, described first message refers to the message not being labeled theme label; According to the topic model of described social networks, obtain the theme label of described first message;
Recommending module, the theme label for the first message by described acquisition is presented on screen and selects for described user;
Determination module, for the result selected according to described user, determines the theme label of described first message.
In conjunction with second aspect, in the first possible embodiment of second aspect, described acquisition module, also for obtaining the second message from described social networks, described second message refers to the message being marked with theme label;
Described device also comprises:
Pretreatment module, for carrying out pre-service to described second message, and preserves described pretreated result;
Study module, for by using machine learning method to train described pretreated result, obtains the topic model of described social networks;
Described acquisition module, specifically for the topic model stating social networks obtained according to described study module, obtains the theme label of described first message.
In conjunction with the first possible embodiment of second aspect, in the embodiment that the second of second aspect is possible, described pretreatment module, specifically for carrying out cutting by described second message by vocabulary; From the result of described cutting, phrase is obtained according to default part of speech; Described phrase is stored according to preset format.
In conjunction with the first possible embodiment of second aspect or the possible embodiment of the second, in the embodiment that the second of second aspect is possible, described study module, specifically for using machine learning method to sample to described pretreated result, obtain theme-vocabulary probability distribution and theme-theme label probability distribution according to described sampled result; According to described theme-vocabulary probability distribution and described theme-theme label probability distribution, obtain the topic model of described social networks.
In conjunction with second aspect to second aspect the third possible embodiment any one of, in the 4th kind of possible embodiment of second aspect, described acquisition module, specifically for adopt formula p ( h | d u ) = Σ t ψ t , h × ( ( Σ r ∈ F u ∪ { U } η u , r θ r , t ) Π w u , d , n ( π 0 φ B , w u , d , n + π 1 φ t , w u , d , n ) ) Calculate described first message d utheme label be the probability of h, and obtain described p (h|d u) be greater than the theme label of the first predetermined threshold value, wherein: ψ t,hrepresent the probability occurring theme label h in the theme label of theme t, η u,rrepresent that user u selects the probability of the history theme of user r, θ r,trepresent that the message of user r issue belongs to the probability of theme t, π 0represent that a vocabulary belongs to the probability of not a theme vocabulary, represent in not a theme vocabulary and occur w u, d, nprobability, π 1represent that a vocabulary belongs to the probability of theme vocabulary, represent to belong in the vocabulary of theme t and occur w u, d, nprobability;
Described recommending module, specifically for by described p (h|d u) the value theme label that is greater than the first predetermined threshold value be presented on screen and select for described user.
In conjunction with the 4th kind of possible embodiment of second aspect, in the 5th kind of possible embodiment of second aspect, described acquisition module, specifically for by the ψ in the topic model of described social networks t,harrange from big to small, and according to ψ t,horder from big to small substitutes into formula successively p ( h | d u ) = Σ t ψ t , h × ( ( Σ r ∈ F u ∪ { U } η u , r θ r , t ) Π w u , d , n ( π 0 φ B , w u , d , n + π 1 φ t , w u , d , n ) ) Calculate, when calculating the p (h|d obtained u) value stops calculating when being less than the second predetermined threshold value, described second predetermined threshold value is less than or equal to described first predetermined threshold value.
In the embodiment of the present invention, obtain the first message not marking theme label that user issues on social networks, according to the topic model of social networks, obtain the theme label of this first message, and these theme label are presented on screen for user's selection, according to the selection result of user, determine the theme label of this first message.Achieve and come to user's proposed topic label by topic model, and finally determine theme label by the selection of user, the final like this theme label determined could think the theme of expression originally closest to user.
Accompanying drawing explanation
In order to be illustrated more clearly in the technical scheme of the embodiment of the present invention, be briefly described to the accompanying drawing used required in embodiment below, apparently, accompanying drawing in the following describes is some embodiments of the present invention, for those of ordinary skill in the art, under the prerequisite not paying creative work, other accompanying drawing can also be obtained according to these accompanying drawings.
Fig. 1 is the schematic flow sheet of theme label recommendation process embodiment of the method one provided by the invention;
Fig. 2 is the schematic flow sheet of theme label recommendation process embodiment of the method two provided by the invention;
Fig. 3 is the structural representation of theme label recommendation process device embodiment one provided by the invention;
Fig. 4 is the structural representation of theme label recommendation process device embodiment two provided by the invention;
Fig. 5 is the structural representation of theme label recommendation process device embodiment three provided by the invention.
Embodiment
For making the object of the embodiment of the present invention, technical scheme and advantage clearly, below in conjunction with the accompanying drawing in the embodiment of the present invention, technical scheme in the embodiment of the present invention is clearly and completely described, obviously, described embodiment is the present invention's part embodiment, instead of whole embodiments.Based on the embodiment in the present invention, those of ordinary skill in the art, not making the every other embodiment obtained under creative work prerequisite, belong to the scope of protection of the invention.
Fig. 1 is the schematic flow sheet of theme label recommendation process embodiment of the method one provided by the invention, and as shown in Figure 1, the method comprises:
The first message that S101, acquisition user issue on social networks, this first message refers to the message not being labeled theme label.
Above-mentioned social networks can be, the microblogging that active user commonly uses, Renren Network, micro-letter circle of friends, push away spy (twitter), the types of facial makeup in Beijing operas (facebook) etc., but not as limit, user can give out information on these social networks, or forward others' message, user likely can add theme label (hashtag) when giving out information, and also may not add hashtag because of trouble and directly issues.The first message herein just refers to the message not adding hashtag.
S102, topic model according to above-mentioned social networks, obtain the theme label of the first message.
This topic model is used for proposed topic label, has broad applicability.
S103, the theme label of the first message of above-mentioned acquisition is presented on screen and selects for above-mentioned user.
Can get multiple theme label for above-mentioned first message according to topic model, these theme label are all shown to user by the display screen of subscriber equipment by these, and user can select wherein, determine oneself to think most suitable theme label.
S104, the result selected according to user, determine the theme label of above-mentioned first message.
User selects one oneself to think most suitable theme label in the theme label of above-mentioned acquisition, and after determining, this theme label is exactly the theme label of the first message.
In the present embodiment, obtain the first message not marking theme label that user issues on social networks, according to the topic model of social networks, obtain the theme label of this first message, and these theme label are presented on screen for user's selection, according to the selection result of user, determine the theme label of this first message.Achieve and come to user's proposed topic label by topic model, and finally determine theme label by the selection of user, the final like this theme label determined could think the theme of expression originally closest to user.Meanwhile, by the first message determination theme label, also facilitate the interested theme of user search or event, help network to carry out the function such as time detecting and monitoring.
Fig. 2 is the schematic flow sheet of theme label recommendation process embodiment of the method two provided by the invention, as shown in Figure 2, before above-mentioned S102, namely according to the topic model of above-mentioned social networks, before obtaining the theme label of the first message, also comprises:
S201, from above-mentioned social networks, obtain the second message, this second message refers to the message being marked with theme label.
While obtaining the second message, the social networks message obtained in social networks can also be comprised.This social networks message can comprise the circle of friends message of user message on social networks and these users.
S202, pre-service is carried out to above-mentioned second message, and preserve this pretreated result.
Particularly, pre-service is carried out to above-mentioned second message, and preserve this pretreated result, Ke Yishi: above-mentioned second message is carried out cutting by vocabulary; From the result of above-mentioned cutting, phrase is obtained according to default part of speech; Above-mentioned phrase is stored according to preset format.
Illustrate, suppose that second message obtained is for " I has spent Christmas Day with several friend together with ", the word combination obtained after carrying out cutting by vocabulary be I with, several, friend, together with, spend, Christmas Day.And wherein " Christmas Day " theme label that is this message.
From the result of above-mentioned cutting, obtain phrase according to default part of speech, can be further above-mentioned word combination is carried out deleting optimization, such as, delete stop words, waste advertisements word etc." several " in above-mentioned word combination, stop words that " " is such are deleted, the word combination after renewal for I with, friend, together with, spend, Christmas Day.
When carrying out storing, the word combination of acquisition and its theme label are stored, and indicate the corresponding relation of vocabulary and theme label.
S203, by using machine learning method to train above-mentioned pretreated result, obtain the topic model of above-mentioned social networks.
When training, the theme label obtained can be trained as existing theme from the second message.
Further, by using machine learning method, above-mentioned pretreated result is trained, obtain the topic model of above-mentioned social networks, be specifically as follows: use machine learning method to sample to above-mentioned pretreated result, obtain theme-vocabulary probability distribution and theme-theme label probability distribution according to above-mentioned sampled result; According to theme-vocabulary probability distribution and theme-theme label probability distribution, obtain the topic model of above-mentioned social networks.
Wherein, theme-vocabulary probability distribution specifically refers to belong to the probability distribution of each vocabulary under same subject; Theme-theme label probability distribution specifically refers to belong to the probability distribution of each theme label under same subject.
In concrete sampling process, above-mentioned pretreated result as sample, can be sampled at random, also can sample according to certain rule.
Using machine learning method to sample to above-mentioned pretreated result, can be carry out gibbs sampler (Gibbs Sampling) to pretreated result.In fact namely the second message is sampled.
But not as limit in the embodiment of the present invention, the variational method (Variational Method) can also be adopted to carry out the study of topic model.
More specifically, be described for concrete certain the second message, pretreated result sampled, mainly comprise three aspects:
(1) the history theme of history theme or the friend of this user that the theme of above-mentioned second message of every bar comes from above-mentioned user is sampled.Namely the theme of a piece of news that user u issues may come from user u once used theme, also may come from certain friend once used theme in user u social networks.Certain sample of such as user u is the theme label of { I, like, recreation ground } is " recreation ground ", and the history theme that sampling analysis " recreation ground " this theme comes from this user u oneself still comes from the history theme of user friend u.
(2) theme of the second message is sampled.What the theme of the second message of namely sampling is.
The theme of the second message may be exactly the theme label of this message, also may be the theme belonging to this theme label." Christmas Day ", " Christmas ", " Christmas ", " silent night " these theme label can belong to " Christmas Day " this theme.
(3) vocabulary in the second message is belonged to theme vocabulary or belongs to not a theme vocabulary sample.I.e. sampling analysis, the vocabulary in certain second message is specifically relevant to theme or irrelevant with theme.Such as " I ", " together ", " spending " this vocabulary that can not embody theme are not a theme vocabulary, and picture " friend ", " Christmas Day " this vocabulary that may become to be the theme are the theme vocabulary.Theme vocabulary can be identical with theme vocabulary, can be also the vocabulary relevant to theme, the vocabulary such as belonging to " Christmas Day " this theme can comprise: " Christmas Day ", " Christmas ", " Christmas ", " silent night " etc. vocabulary.
It should be noted that, when specifically performing, above-mentioned (1), (2), (3) are order in no particular order, can perform simultaneously, also can perform in any order, be not construed as limiting in the embodiment of the present invention.
More specifically, for above-mentioned (1) aspect, sample to the history theme of history theme or the friend of this user that the theme of the second message comes from above-mentioned user, Ke Yishi, adopts formula p ( f u , d = r | f ⫬ ( u , d ) , z , y , w , h ) ∝ cf u , r , ⫬ ( u , d ) + δ cf u , ( · ) , ⫬ ( u , d ) + ( | F u | + 1 ) δ × ct r , z u , d , ⫬ ( u , d ) + α ct r , ( · ) , ⫬ ( u , d ) + ( | T | ) α (being designated as formula 1) history theme to history theme or the friend of this user that the theme of the second message comes from above-mentioned user is sampled.History theme refers to the used theme of user.Wherein: represent that the theme of above-mentioned second message comes from the probability of the history theme of user r, user r is the friend of above-mentioned user (being designated as: user u) oneself or user u; | F u| represent friend's sum of user u; represent the number of times employing the history theme of user r in the message that user u issues except above-mentioned second message; represent and own except above-mentioned second message the summation of value, namely except above-mentioned second message, for all different user r's the summation of value, namely substitutes into that all different r values obtain the summation of value; z u,drepresent the theme of this second message; what expression described user r except above-mentioned second message issued belongs to described theme z u,dthe quantity of message; represent the sum of the message that user r issues except above-mentioned second message; | T| represents the current quantity that there is theme.These parameters substituting into formula can be obtained by statistics.
For above-mentioned (2) aspect, sample to the theme of above-mentioned second message, Ke Yishi, adopts formula:
p ( z u , d = i | z ⫬ ( u , d ) , f , y , w , h ) ∝ ct f u , d , i , ⫬ ( u , d ) + α ct f u , d , ( · ) , ⫬ ( u , d ) + ( | T | ) α
× ( Π k = 0 nw i , ( · ) - 1 1 cw i , ( · ) , ⫬ ( u , d ) + | V | β + k ) (being designated as formula 2)
× ( Π v ∈ w u , d Π k = 0 nw i , w - 1 ( cw i , v , ⫬ ( u , d ) + β + k ) )
× ( Π k = 0 M u , d - 1 1 ch i , ( · ) , ⫬ ( u , d ) + | H | μ + k ) ( Π g ∈ h u , d ( ch i , g , ⫬ ( u , d ) + μ ) )
The theme of above-mentioned second message is sampled, wherein: represent that above-mentioned second message subject is the theme the probability of i; represent the quantity belonging to the message of theme i in the message that above-mentioned user issues except above-mentioned second message; represent the sum of message except above-mentioned second message that above-mentioned user issues, such as user has issued 100 message, then be 99; represent that vocabulary v(vocabulary v represents the vocabulary that in above-mentioned second message one is concrete except above-mentioned second message) belong to the number of times of the vocabulary of theme i; represent the sum belonging to the vocabulary of theme i except above-mentioned second message; w u,drepresent all vocabulary in above-mentioned second message, vocabulary v just represents w u,done of them; Nw i,vrepresent that vocabulary v belongs to the number of times of the vocabulary of theme i; | V| represents the current quantity that there is vocabulary; | H| represents the current quantity that there is theme label; represent and except above-mentioned second message, include theme label g and the quantity belonging to the message of theme i; represent all the summation of value, namely includes the sum of the message of theme label g except above-mentioned second message.These parameters substituting into formula can be obtained by statistics.
For above-mentioned (3) aspect, the vocabulary in above-mentioned second message is belonged to theme vocabulary or belongs to not a theme vocabulary and samples, Ke Yishi, adopt formula p ( y u , d , n = 1 | y ⫬ ( u , d , n ) , f , z , w , h ) ∝ cy 1 , ⫬ ( u , d , n ) + γ cy ( · ) , ⫬ ( u , d , n ) + 2 γ × cw i , w u , d , n , ⫬ ( u , d , n ) + β cw i , ( · ) , ⫬ ( u , d , n ) + | V | β (being designated as formula 3) belongs to theme vocabulary (non-background vocabulary) to the vocabulary in above-mentioned second message and samples.Adopt formula p ( y u , d , n = 0 | y ⫬ ( u , d , n ) , f , z , w , h ) ∝ cy 0 , ⫬ ( u , d , n ) + γ cy ( · ) , ⫬ ( u , d , n ) + 2 γ × cw B , w u , d , n , ⫬ ( u , d , n ) + β cw B , ( · ) , ⫬ ( u , d , n ) + | V | β (being designated as formula 4) belongs to theme vocabulary (background vocabulary) to the vocabulary in above-mentioned second message and samples.
Wherein: the vocabulary n(vocabulary n representing in above-mentioned second message represents the vocabulary that in above-mentioned second message is concrete) belong to the probability of theme vocabulary, represent that the vocabulary n in above-mentioned second message belongs to the probability of not a theme vocabulary; represent the vocabulary quantity belonging to theme vocabulary except the vocabulary n in described second message; represent the sum of vocabulary except the vocabulary n in the second message; represent the vocabulary quantity belonging to not a theme vocabulary except the vocabulary n in described second message; represent w u, d, nthe number of times of theme i is being appeared at, wherein w except the vocabulary n in described second message u,drepresent all vocabulary in above-mentioned second message, vocabulary n just represents w u,done of them; represent the sum of different vocabulary there is theme i except the vocabulary n in described second message under, | V| represents the current quantity that there is vocabulary.These parameters substituting into formula can be obtained by statistics.
According to the result of above-mentioned Gibbs Sampling, the topic model got comprises following parameter: 1) vocabulary belongs to the probability of not a theme vocabulary or belongs to the probability of theme vocabulary, is designated as π k, k=0,1, represent that as k=0 this vocabulary is not a theme vocabulary, during k=1, represent that this vocabulary is the theme vocabulary.2) message that above-mentioned user (being designated as user u) issues belongs to the probability of theme t, is designated as θ u,t.3) in not a theme vocabulary, there is the probability of vocabulary v, be designated as φ b,v.4) in the vocabulary belonging to theme t, there is the probability of vocabulary v, be designated as φ t,v.5) in the theme label of theme t, there is the probability of theme label h, be designated as ψ t,h.6) above-mentioned user (user u) selects the probability of the history theme of user r, is designated as η u,r, user r is the friend of user u or user u itself, and in the embodiment of the present invention, friend all refers to the friend of user on social networks.
α, β, γ, μ, δ in above-mentioned sampling are respectively the Di Li Cray priori (Dirichlet prior) of parameter θ, φ, π, ψ, η in above-mentioned topic model.
More specifically, can, according to the result of Gibbs Sampling, following formula be adopted to calculate above-mentioned parameter respectively: 1) k=0,1(are designated as formula 5); 2) t ∈ T(is designated as formula 6); 3) (being designated as formula 7); 4) φ t , v = cw t , v + β cw t , ( · ) + β , ∀ t ∈ T , V ∈ V(is designated as formula 8); 5) ψ t , h = ch t , h + μ ch t , ( · ) + | H | μ , ∀ t ∈ T , H ∈ H(is designated as formula 8); 6) η u , r = cF u , r + δ cF u , ( · ) + ( | F u | + 1 ) δ , ∀ u ∈ U , r ∈ | F u | ∪ { u } (being designated as formula 8).
Wherein, U represents the set of all users crawled; Cy krepresent and belong to the quantity of not a theme vocabulary or the quantity of theme vocabulary, wherein represent the quantity of not a theme vocabulary during k=0, during k=1, represent the quantity of theme vocabulary; Cy ()represent the summation of the quantity of theme vocabulary and the quantity of not a theme vocabulary; Ct u,trepresent the quantity belonging to the message of theme t that user u issues; Ct u, ()represent the sum of the message that user u issues; Cw b,vrepresent that vocabulary v is the number of times of not a theme vocabulary; Cw b, ()represent the summation of not a theme vocabulary occurrence number; Cw t,vrepresent that vocabulary v belongs to the number of times of theme t; Cw t, ()represent the summation of the number of times that the vocabulary belonging to theme t occurs; Ch t,hrepresent the number of times occurring theme label h in the theme label of theme t; Ch t, ()represent the sum of all theme label of theme t; Cf u,rrepresent that user u uses the number of times of the history theme of user r; Cf u, ()represent the sum of the used theme of user u.These parameters come from aforementioned sample result and statistics.
It should be noted that, in order to improve operation efficiency, the process of above-mentioned acquisition topic model can be processed by multiple parallel processor simultaneously, is finally arranged by keeper (manager) again, exports the topic model got.
Certainly, in specific implementation process, constantly can swash from social networks according to the time interval of presetting and fetch data, also above-mentioned flow process can constantly be repeated, for the new data obtained, if just get the message that user newly issues, the theme label wherein adopted currently there is label, so this message performs above-mentioned sampling as new sample, and upgrades topic model; If the new information got, further comprises new theme label, so first on the vector dimension of current already present theme label, add 1, then above-mentioned sampling is carried out to this sample, and upgrade topic model.
Further, in order to improve the efficiency that theme label is recommended, can also be optimized topic model.In specific implementation process, after obtaining topic model, can will be less than the ψ of the 3rd predetermined threshold value in above-mentioned topic model on the one hand t,hcorresponding theme label from current existed theme label remove, namely preset the 3rd predetermined threshold value and existing ψ t,hcompare, will the ψ of the 3rd predetermined threshold value be less than t,hcorresponding theme label from stored current existed theme label remove, just need not consider the theme label removed during such subsequent recommendation theme label, reduce operand, raise the efficiency.Similarly, on the other hand, the θ of the 4th predetermined threshold value will can be less than in above-mentioned topic model u,tcorresponding theme from current existed theme remove, be exactly without the theme that these remove in subsequent recommendation process.Particularly, in specific implementation process, the optimization of above-mentioned two aspects can be performed simultaneously, also only can perform one of them as the case may be.
Further, the above-mentioned topic model according to social networks, obtains the theme label of above-mentioned first message, is specially employing formula p ( h | d u ) = Σ t ψ t , h × ( ( Σ r ∈ F u ∪ { U } η u , r θ r , t ) Π w u , d , n ( π 0 φ B , w u , d , n + π 1 φ t , w u , d , n ) ) Calculate described first message d utheme label be the probability of h, and obtain described p (h|d u) value be greater than the theme label of the first predetermined threshold value.Wherein: ψ t,hrepresent the probability occurring theme label h in the theme label of theme t, η u,rrepresent that above-mentioned user (user u) selects the probability of the history theme of user r, θ r,trepresent that the message of user r issue belongs to the probability of theme t, π 0represent that a vocabulary belongs to the probability of not a theme vocabulary, represent in not a theme vocabulary and occur w u, d, nprobability, π 1represent that a vocabulary belongs to the probability of theme vocabulary, represent to belong in the vocabulary of theme t and occur w u, d, nprobability, w u, d, nrepresent the vocabulary n in described first message.W u, d, nrepresent the vocabulary n in described first message.
Correspondingly, the above-mentioned theme label by the first message obtained is presented on screen and selects for user, is specially: by above-mentioned p (h|d u) the value theme label that is greater than the first predetermined threshold value be presented on screen and select for described user.
In another kind of mode, also can be, by the p (h|d calculated u) value arranges, from big to small in order by p (h|d u) maximum N number of theme label be presented on screen for user select, wherein N is the number of default proposed topic label.
Theme label h be current any one that existed in theme label, substitute into these values successively for different theme label and carry out calculating p (h|d u) value.Owing to being independently for the computing of each theme label h, during specific implementation, in order to improve operand, different theme label can be processed by multiple processor simultaneously, obtaining p (h|d u) value, finally carry out arranging by keeper (manager) theme label obtaining being shown to user again.
Further, in order to improve the efficiency of proposed topic label, above-mentioned employing formula p ( h | d u ) = Σ t ψ t , h × ( ( Σ r ∈ F u ∪ { U } η u , r θ r , t ) Π w u , d , n ( π 0 φ B , w u , d , n + π 1 φ t , w u , d , n ) ) Calculate above-mentioned first message d utheme label be the probability of h, Ke Yishi, by the ψ in the topic model of above-mentioned social networks t,harrange from big to small, and according to ψ t,horder from big to small substitutes into formula successively p ( h | d u ) = Σ t ψ t , h × ( ( Σ r ∈ F u ∪ { U } η u , r θ r , t ) Π w u , d , n ( π 0 φ B , w u , d , n + π 1 φ t , w u , d , n ) ) Calculate, when calculating the p (h|d obtained u) value stops calculating when being less than the second predetermined threshold value, described second predetermined threshold value is less than or equal to described first predetermined threshold value.Preset second predetermined threshold value like this, just unnecessary all parameters substitution calculates, and reduces operand, improves work efficiency.
Fig. 3 is the structural representation of theme label recommendation process device embodiment one provided by the invention, and as shown in Figure 3, this device comprises: acquisition module 301, recommending module 302 and determination module 304, wherein:
Acquisition module 301, for obtaining the first message that user issues on social networks, described first message refers to the message not being labeled theme label; According to the topic model of described social networks, obtain the theme label of described first message.
Recommending module 302, the theme label for the first message by described acquisition is presented on screen and selects for described user.
Determination module 303, for the result selected according to described user, determines the theme label of described first message.
This device is for performing preceding method embodiment, and it realizes principle and technique effect is similar, does not repeat them here.
Fig. 4 is the structural representation of theme label recommendation process device embodiment two provided by the invention, and as shown in Figure 4, on the basis of Fig. 3, this device can also comprise: pretreatment module 401 and study module 402.
In specific implementation process:
Above-mentioned acquisition module 301, also for obtaining the second message from described social networks, described second message refers to the message being marked with theme label.
Pretreatment module 401, for carrying out pre-service to described second message, and preserves described pretreated result.
Study module 402, for by using machine learning method to train described pretreated result, obtains the topic model of described social networks.
Acquisition module 301, specifically for the topic model stating social networks obtained according to study module 402, obtains the theme label of described first message.
Further, pretreatment module 401, specifically for carrying out cutting by described second message by vocabulary; From the result of described cutting, phrase is obtained according to default part of speech; Described phrase is stored according to preset format.
Study module 402, specifically for using machine learning method to sample to described pretreated result, obtains theme-vocabulary probability distribution and theme-theme label probability distribution according to described sampled result; According to described theme-vocabulary probability distribution and described theme-theme label probability distribution, obtain the topic model of described social networks.
Further, above-mentioned acquisition module 301, specifically for adopting formula p ( h | d u ) = Σ t ψ t , h × ( ( Σ r ∈ F u ∪ { U } η u , r θ r , t ) Π w u , d , n ( π 0 φ B , w u , d , n + π 1 φ t , w u , d , n ) ) Calculate described first message d utheme label be the probability of h, and obtain described p (h|d u) value be greater than the theme label of the first predetermined threshold value, wherein: ψ t,hrepresent the probability occurring theme label h in the theme label of theme t, η u,rrepresent that described user selects the probability of the history theme of user r, θ r,trepresent that the message of user r issue belongs to the probability of theme t, π 0represent that a vocabulary belongs to the probability of not a theme vocabulary, represent in not a theme vocabulary and occur w u, d, nprobability, π 1represent that a vocabulary belongs to the probability of theme vocabulary, represent to belong in the vocabulary of theme t and occur w u, d, nprobability, w u, d, nrepresent the vocabulary n in described first message.
Recommending module 302, specifically for by described p (h|d u) the value theme label that is greater than the first predetermined threshold value be presented on screen and select for described user.
Alternatively, acquisition module 301, specifically for by the ψ in the topic model of described social networks t,harrange from big to small, and according to ψ t,horder from big to small substitutes into formula successively p ( h | d u ) = Σ t ψ t , h × ( ( Σ r ∈ F u ∪ { U } η u , r θ r , t ) Π w u , d , n ( π 0 φ B , w u , d , n + π 1 φ t , w u , d , n ) ) Calculate, when calculating the p (h|d obtained u) value stops calculating when being less than the second predetermined threshold value, described second predetermined threshold value is less than or equal to described first predetermined threshold value.
This device is for performing preceding method embodiment, and it realizes principle and technique effect is similar, does not repeat them here.
Fig. 5 is the structural representation of theme label recommendation process device embodiment three provided by the invention, and as shown in Figure 5, this device comprises: storer 501 and processor 502.Wherein:
Storer 501 is for store sets of instructions.This processor 502 is configured to call the instruction set in storer 501, to perform following flow process: obtain the first message that user issues on social networks, described first message refers to the message not being labeled theme label; According to the topic model of described social networks, obtain the theme label of described first message; The theme label of the first message of described acquisition is presented on screen and selects for described user; According to the result that described user selects, determine the theme label of described first message.
This device is for performing preceding method embodiment, and it realizes principle and technique effect is similar, does not repeat them here.
Processor 502, also for obtaining the second message from described social networks, described second message refers to the message being marked with theme label; Pre-service is carried out to described second message, and preserves described pretreated result; By using machine learning method to train described pretreated result, obtain the topic model of described social networks.
Further, processor 502, specifically for carrying out cutting by described second message by vocabulary; From the result of described cutting, phrase is obtained according to default part of speech; Described phrase is stored according to preset format.
Processor 502, specifically for using machine learning method to sample to described pretreated result, obtains theme-vocabulary probability distribution and theme-theme label probability distribution according to described sampled result; According to described theme-vocabulary probability distribution and described theme-theme label probability distribution, obtain the topic model of described social networks.
Processor 502, specifically for adopting formula
p ( h | d u ) = Σ t ψ t , h × ( ( Σ r ∈ F u ∪ { U } η u , r θ r , t ) Π w u , d , n ( π 0 φ B , w u , d , n + π 1 φ t , w u , d , n ) ) Calculate described first message d utheme label be the probability of h, and obtain described p (h|d u) value be greater than the theme label of the first predetermined threshold value, wherein: ψ t,hrepresent the probability occurring theme label h in the theme label of theme t, η u,rrepresent that user u selects the probability of the history theme of user r, θ r,trepresent that the message of user r issue belongs to the probability of theme t, π 0represent that a vocabulary belongs to the probability of not a theme vocabulary, represent in not a theme vocabulary and occur w u, d, nprobability, π 1represent that a vocabulary belongs to the probability of theme vocabulary, represent with the vocabulary belonging to theme t in there is w u, d, nprobability.
Processor 502, specifically for by described p (h|d u) the value theme label that is greater than the first predetermined threshold value be presented on screen and select for described user.
Alternatively, processor 502, specifically for by the ψ in the topic model of described social networks t,harrange from big to small, and according to ψ t,horder from big to small substitutes into formula successively p ( h | d u ) = Σ t ψ t , h × ( ( Σ r ∈ F u ∪ { U } η u , r θ r , t ) Π w u , d , n ( π 0 φ B , w u , d , n + π 1 φ t , w u , d , n ) ) Calculate, when calculating the p (h|d obtained u) value stops calculating when being less than the second predetermined threshold value, described second predetermined threshold value is less than or equal to described first predetermined threshold value.
In several embodiment provided by the present invention, should be understood that, disclosed apparatus and method, can realize by another way.Such as, device embodiment described above is only schematic, such as, the division of described unit, be only a kind of logic function to divide, actual can have other dividing mode when realizing, such as multiple unit or assembly can in conjunction with or another system can be integrated into, or some features can be ignored, or do not perform.Another point, shown or discussed coupling each other or direct-coupling or communication connection can be by some interfaces, and the indirect coupling of device or unit or communication connection can be electrical, machinery or other form.
The described unit illustrated as separating component or can may not be and physically separates, and the parts as unit display can be or may not be physical location, namely can be positioned at a place, or also can be distributed in multiple network element.Some or all of unit wherein can be selected according to the actual needs to realize the object of the present embodiment scheme.
In addition, each functional unit in each embodiment of the present invention can be integrated in a processing unit, also can be that the independent physics of unit exists, also can two or more unit in a unit integrated.Above-mentioned integrated unit both can adopt the form of hardware to realize, and the form that hardware also can be adopted to add SFU software functional unit realizes.
The above-mentioned integrated unit realized with the form of SFU software functional unit, can be stored in a computer read/write memory medium.Above-mentioned SFU software functional unit is stored in a storage medium, comprising some instructions in order to make a computer equipment (can be personal computer, server, or the network equipment etc.) or processor (processor) perform the part steps of method described in each embodiment of the present invention.And aforesaid storage medium comprises: USB flash disk, portable hard drive, ROM (read-only memory) (Read-Only Memory, ROM), random access memory (Random Access Memory, RAM), magnetic disc or CD etc. various can be program code stored medium.
Last it is noted that above each embodiment is only in order to illustrate technical scheme of the present invention, be not intended to limit; Although with reference to foregoing embodiments to invention has been detailed description, those of ordinary skill in the art is to be understood that: it still can be modified to the technical scheme described in foregoing embodiments, or carries out equivalent replacement to wherein some or all of technical characteristic; And these amendments or replacement, do not make the essence of appropriate technical solution depart from the scope of various embodiments of the present invention technical scheme.

Claims (12)

1. a theme label recommend method, is characterized in that, comprising:
Obtain the first message that user issues on social networks, described first message refers to the message not being labeled theme label;
According to the topic model of described social networks, obtain the theme label of described first message;
The theme label of the first message of described acquisition is presented on screen and selects for described user;
According to the result that described user selects, determine the theme label of described first message.
2. method according to claim 1, is characterized in that, the described topic model according to described social networks, before obtaining the theme label of described first message, also comprises:
From described social networks, obtain the second message, described second message refers to the message being marked with theme label;
Pre-service is carried out to described second message, and preserves described pretreated result;
By using machine learning method to train described pretreated result, obtain the topic model of described social networks.
3. method according to claim 2, is characterized in that, describedly carries out pre-service to described second message, and preserves described pretreated result, comprising:
Described second message is carried out cutting by vocabulary;
From the result of described cutting, phrase is obtained according to default part of speech;
Described phrase is stored according to preset format.
4. according to the method in claim 2 or 3, it is characterized in that, described by using machine learning method to train described pretreated result, the topic model obtaining described social networks comprises:
Use machine learning method to sample to described pretreated result, obtain theme-vocabulary probability distribution and theme-theme label probability distribution according to described sampled result;
According to described theme-vocabulary probability distribution and described theme-theme label probability distribution, obtain the topic model of described social networks.
5. the method according to any one of Claims 1-4, is characterized in that, the described topic model according to described social networks, and the theme label obtaining described first message comprises:
Adopt formula p ( h | d u ) = Σ t ψ t , h × ( ( Σ r ∈ F u ∪ { U } η u , r θ r , t ) Π w u , d , n ( π 0 φ B , w u , d , n + π 1 φ t , w u , d , n ) ) Calculate described first message d utheme label be the probability of h, and obtain described p (h|d u) value be greater than the theme label of the first predetermined threshold value, wherein: ψ t,hrepresent the probability occurring theme label h in the theme label of theme t, η u,rrepresent that described user selects the probability of the history theme of user r, θ r,trepresent that the message of user r issue belongs to the probability of theme t, π 0represent that a vocabulary belongs to the probability of not a theme vocabulary, represent in not a theme vocabulary and occur w u, d, nprobability, π 1represent that a vocabulary belongs to the probability of theme vocabulary, represent to belong in the vocabulary of theme t and occur w u, d, nprobability, w u, d, nrepresent the vocabulary n in described first message;
The described theme label by the first message of described acquisition is presented on screen and selects to comprise for described user:
By described p (h|d u) the value theme label that is greater than the first predetermined threshold value be presented on screen and select for described user.
6. method according to claim 5, is characterized in that, described employing formula p ( h | d u ) = Σ t ψ t , h × ( ( Σ r ∈ F u ∪ { U } η u , r θ r , t ) Π w u , d , n ( π 0 φ B , w u , d , n + π 1 φ t , w u , d , n ) ) Calculate described first message d utheme label be the probability of h, comprising:
By the ψ in the topic model of described social networks t,harrange from big to small, and according to ψ t,horder from big to small substitutes into formula successively p ( h | d u ) = Σ t ψ t , h × ( ( Σ r ∈ F u ∪ { U } η u , r θ r , t ) Π w u , d , n ( π 0 φ B , w u , d , n + π 1 φ t , w u , d , n ) ) Calculate, when calculating the p (h|d obtained u) value stops calculating when being less than the second predetermined threshold value, described second predetermined threshold value is less than or equal to described first predetermined threshold value.
7. a theme label recommendation apparatus, is characterized in that, comprising:
Acquisition module, for obtaining the first message that user issues on social networks, described first message refers to the message not being labeled theme label; According to the topic model of described social networks, obtain the theme label of described first message;
Recommending module, the theme label for the first message by described acquisition is presented on screen and selects for described user;
Determination module, for the result selected according to described user, determines the theme label of described first message.
8. device according to claim 7, is characterized in that, described acquisition module, and also for obtaining the second message from described social networks, described second message refers to the message being marked with theme label;
Described device also comprises:
Pretreatment module, for carrying out pre-service to described second message, and preserves described pretreated result;
Study module, for by using machine learning method to train described pretreated result, obtains the topic model of described social networks;
Described acquisition module, specifically for the topic model stating social networks obtained according to described study module, obtains the theme label of described first message.
9. device according to claim 8, is characterized in that, described pretreatment module, specifically for described second message is carried out cutting by vocabulary; From the result of described cutting, phrase is obtained according to default part of speech; Described phrase is stored according to preset format.
10. device according to claim 8 or claim 9, it is characterized in that, described study module, specifically for using machine learning method to sample to described pretreated result, obtains theme-vocabulary probability distribution and theme-theme label probability distribution according to described sampled result; According to described theme-vocabulary probability distribution and described theme-theme label probability distribution, obtain the topic model of described social networks.
11. devices according to any one of claim 7 to 10, is characterized in that, described acquisition module, specifically for adopting formula p ( h | d u ) = Σ t ψ t , h × ( ( Σ r ∈ F u ∪ { U } η u , r θ r , t ) Π w u , d , n ( π 0 φ B , w u , d , n + π 1 φ t , w u , d , n ) ) Calculate described first message d utheme label be the probability of h, and obtain described p (h|d u) be greater than the theme label of the first predetermined threshold value, wherein: ψ t,hrepresent the probability occurring theme label h in the theme label of theme t, η u,rrepresent that user u selects the probability of the history theme of user r, θ r,trepresent that the message of user r issue belongs to the probability of theme t, π 0represent that a vocabulary belongs to the probability of not a theme vocabulary, represent in not a theme vocabulary and occur w u, d, nprobability, π 1represent that a vocabulary belongs to the probability of theme vocabulary, represent to belong in the vocabulary of theme t and occur w u, d, nprobability;
Described recommending module, specifically for by described p (h|d u) the value theme label that is greater than the first predetermined threshold value be presented on screen and select for described user.
12. devices according to claim 7, is characterized in that, described acquisition module, specifically for by the ψ in the topic model of described social networks t,harrange from big to small, and according to ψ t,horder from big to small substitutes into formula successively p ( h | d u ) = Σ t ψ t , h × ( ( Σ r ∈ F u ∪ { U } η u , r θ r , t ) Π w u , d , n ( π 0 φ B , w u , d , n + π 1 φ t , w u , d , n ) ) Calculate, when calculating the p (h|d obtained u) value stops calculating when being less than the second predetermined threshold value, described second predetermined threshold value is less than or equal to described first predetermined threshold value.
CN201410096024.1A 2014-03-14 2014-03-14 Theme label recommended method and device Active CN104915359B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201410096024.1A CN104915359B (en) 2014-03-14 2014-03-14 Theme label recommended method and device

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201410096024.1A CN104915359B (en) 2014-03-14 2014-03-14 Theme label recommended method and device

Publications (2)

Publication Number Publication Date
CN104915359A true CN104915359A (en) 2015-09-16
CN104915359B CN104915359B (en) 2019-05-28

Family

ID=54084427

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201410096024.1A Active CN104915359B (en) 2014-03-14 2014-03-14 Theme label recommended method and device

Country Status (1)

Country Link
CN (1) CN104915359B (en)

Cited By (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN106682008A (en) * 2015-11-06 2017-05-17 北京奇虎科技有限公司 Friend circle information classification method and device
CN107145553A (en) * 2017-04-28 2017-09-08 暴风集团股份有限公司 A kind of Network Data Capture method and system for competitive sports
CN107423308A (en) * 2016-05-24 2017-12-01 华为技术有限公司 subject recommending method and device
CN108335165A (en) * 2017-02-10 2018-07-27 腾讯科技(深圳)有限公司 Interest tags determine method and apparatus
CN110084710A (en) * 2018-01-26 2019-08-02 深圳市腾讯计算机系统有限公司 Determine the method and device of message subject
CN111104545A (en) * 2018-10-26 2020-05-05 阿里巴巴集团控股有限公司 Background music configuration method and equipment, client device and electronic equipment
CN111191141A (en) * 2020-01-06 2020-05-22 支付宝(杭州)信息技术有限公司 Document recommendation method and device
CN111400517A (en) * 2020-03-20 2020-07-10 北京字节跳动网络技术有限公司 Information pushing and information publishing method and device

Citations (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101196904A (en) * 2007-11-09 2008-06-11 清华大学 News keyword abstraction method based on word frequency and multi-component grammar
CN101464898A (en) * 2009-01-12 2009-06-24 腾讯科技(深圳)有限公司 Method for extracting feature word of text
CN102624627A (en) * 2012-03-15 2012-08-01 杭州华三通信技术有限公司 Message transmission method and equipment
CN102637165A (en) * 2012-02-17 2012-08-15 清华大学 Method for extracting attribute-viewpoint pairs of Chinese viewpoint and evaluation information
CN102646134A (en) * 2012-03-29 2012-08-22 百度在线网络技术(北京)有限公司 Method and device for determining message session in message record
CN102867016A (en) * 2012-07-18 2013-01-09 北京开心人信息技术有限公司 Label-based social network user interest mining method and device
US20130254231A1 (en) * 2012-03-20 2013-09-26 Kawf.Com, Inc. Dba Tagboard.Com Gathering and contributing content across diverse sources
CN103544255A (en) * 2013-10-15 2014-01-29 常州大学 Text semantic relativity based network public opinion information analysis method

Patent Citations (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101196904A (en) * 2007-11-09 2008-06-11 清华大学 News keyword abstraction method based on word frequency and multi-component grammar
CN101464898A (en) * 2009-01-12 2009-06-24 腾讯科技(深圳)有限公司 Method for extracting feature word of text
CN102637165A (en) * 2012-02-17 2012-08-15 清华大学 Method for extracting attribute-viewpoint pairs of Chinese viewpoint and evaluation information
CN102624627A (en) * 2012-03-15 2012-08-01 杭州华三通信技术有限公司 Message transmission method and equipment
US20130254231A1 (en) * 2012-03-20 2013-09-26 Kawf.Com, Inc. Dba Tagboard.Com Gathering and contributing content across diverse sources
CN102646134A (en) * 2012-03-29 2012-08-22 百度在线网络技术(北京)有限公司 Method and device for determining message session in message record
CN102867016A (en) * 2012-07-18 2013-01-09 北京开心人信息技术有限公司 Label-based social network user interest mining method and device
CN103544255A (en) * 2013-10-15 2014-01-29 常州大学 Text semantic relativity based network public opinion information analysis method

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
FRÉDERIC GODIN ETC: "Using Topic Models for Twitter Hashtag Recommendation", 《NTERNATIONAL WORLD WIDE WEB CONFERENCES STEERING COMMITTEE》 *

Cited By (15)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN106682008A (en) * 2015-11-06 2017-05-17 北京奇虎科技有限公司 Friend circle information classification method and device
CN107423308A (en) * 2016-05-24 2017-12-01 华为技术有限公司 subject recommending method and device
US20190087884A1 (en) 2016-05-24 2019-03-21 Huawei Technologies Co., Ltd. Theme recommendation method and apparatus
US11830033B2 (en) 2016-05-24 2023-11-28 Huawei Technologies Co., Ltd. Theme recommendation method and apparatus
CN107423308B (en) * 2016-05-24 2020-07-07 华为技术有限公司 Theme recommendation method and device
US11226991B2 (en) 2017-02-10 2022-01-18 Tencent Technology (Shenzhen) Company Limited Interest tag determining method, computer device, and storage medium
CN108335165A (en) * 2017-02-10 2018-07-27 腾讯科技(深圳)有限公司 Interest tags determine method and apparatus
CN108335165B (en) * 2017-02-10 2021-12-14 腾讯科技(深圳)有限公司 Interest tag determination method and device
CN107145553A (en) * 2017-04-28 2017-09-08 暴风集团股份有限公司 A kind of Network Data Capture method and system for competitive sports
CN110084710A (en) * 2018-01-26 2019-08-02 深圳市腾讯计算机系统有限公司 Determine the method and device of message subject
CN110084710B (en) * 2018-01-26 2022-12-09 深圳市腾讯计算机系统有限公司 Method and device for determining message theme
CN111104545A (en) * 2018-10-26 2020-05-05 阿里巴巴集团控股有限公司 Background music configuration method and equipment, client device and electronic equipment
CN111191141B (en) * 2020-01-06 2023-05-23 蚂蚁胜信(上海)信息技术有限公司 Text recommendation method and device
CN111191141A (en) * 2020-01-06 2020-05-22 支付宝(杭州)信息技术有限公司 Document recommendation method and device
CN111400517A (en) * 2020-03-20 2020-07-10 北京字节跳动网络技术有限公司 Information pushing and information publishing method and device

Also Published As

Publication number Publication date
CN104915359B (en) 2019-05-28

Similar Documents

Publication Publication Date Title
CN104915359A (en) Theme label recommending method and device
CN108334533B (en) Keyword extraction method and device, storage medium and electronic device
US10546005B2 (en) Perspective data analysis and management
CN107346336B (en) Information processing method and device based on artificial intelligence
US20180181988A1 (en) Method and apparatus for pushing information
CN107193974B (en) Regional information determination method and device based on artificial intelligence
CN107146112B (en) Mobile internet advertisement putting method
CN104866478B (en) Malicious text detection and identification method and device
CN103207899A (en) Method and system for recommending text files
US8775423B2 (en) Data mining across multiple social platforms
CN108305180B (en) Friend recommendation method and device
CN107908662B (en) Method and device for realizing search system
US11640420B2 (en) System and method for automatic summarization of content with event based analysis
US20180330278A1 (en) Processes and techniques for more effectively training machine learning models for topically-relevant two-way engagement with content consumers
CN105989107A (en) Application recommendation method and device
CN113657113A (en) Text processing method and device and electronic equipment
CN112667802A (en) Service information input method, device, server and storage medium
US20150379112A1 (en) Creating an on-line job function ontology
EP3009942A1 (en) Social contact message monitoring method and device
US10042913B2 (en) Perspective data analysis and management
CN112148841B (en) Object classification and classification model construction method and device
CN113590756A (en) Information sequence generation method and device, terminal equipment and computer readable medium
CN109710634B (en) Method and device for generating information
CN110516162A (en) A kind of information recommendation method, device, equipment and storage medium
CN108549722A (en) Multi-platform data publication method, system and medium

Legal Events

Date Code Title Description
C06 Publication
PB01 Publication
C10 Entry into substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant
TR01 Transfer of patent right

Effective date of registration: 20210421

Address after: Unit 3401, unit a, building 6, Shenye Zhongcheng, No. 8089, Hongli West Road, Donghai community, Xiangmihu street, Futian District, Shenzhen, Guangdong 518040

Patentee after: Honor Device Co.,Ltd.

Address before: 518129 Bantian HUAWEI headquarters office building, Longgang District, Guangdong, Shenzhen

Patentee before: HUAWEI TECHNOLOGIES Co.,Ltd.

TR01 Transfer of patent right