CN111294619B - Long-short term interest modeling method for IPTV field - Google Patents
Long-short term interest modeling method for IPTV field Download PDFInfo
- Publication number
- CN111294619B CN111294619B CN202010129277.XA CN202010129277A CN111294619B CN 111294619 B CN111294619 B CN 111294619B CN 202010129277 A CN202010129277 A CN 202010129277A CN 111294619 B CN111294619 B CN 111294619B
- Authority
- CN
- China
- Prior art keywords
- term
- long
- short
- click sequence
- attention
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Expired - Fee Related
Links
- 238000000034 method Methods 0.000 title claims abstract description 26
- 230000007774 longterm Effects 0.000 claims abstract description 96
- 238000013507 mapping Methods 0.000 claims abstract description 23
- 238000004364 calculation method Methods 0.000 claims description 16
- 239000011159 matrix material Substances 0.000 claims description 11
- 238000011176 pooling Methods 0.000 claims description 9
- 230000004927 fusion Effects 0.000 claims description 8
- 238000010606 normalization Methods 0.000 claims description 8
- 230000007246 mechanism Effects 0.000 claims description 7
- 230000009466 transformation Effects 0.000 claims description 6
- 239000004576 sand Substances 0.000 claims description 5
- 239000013598 vector Substances 0.000 claims description 5
- 230000006870 function Effects 0.000 claims description 3
- 230000004913 activation Effects 0.000 claims description 2
- 230000008859 change Effects 0.000 claims description 2
- 230000014509 gene expression Effects 0.000 claims 1
- 230000008569 process Effects 0.000 description 5
- 230000009901 attention process Effects 0.000 description 1
- 230000009286 beneficial effect Effects 0.000 description 1
- 244000309464 bull Species 0.000 description 1
- 238000013135 deep learning Methods 0.000 description 1
- 238000010586 diagram Methods 0.000 description 1
- 230000002452 interceptive effect Effects 0.000 description 1
- 230000004048 modification Effects 0.000 description 1
- 238000012986 modification Methods 0.000 description 1
- 238000003672 processing method Methods 0.000 description 1
- 230000009467 reduction Effects 0.000 description 1
- 238000012163 sequencing technique Methods 0.000 description 1
- 239000004575 stone Substances 0.000 description 1
Images
Classifications
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/20—Servers specifically adapted for the distribution of content, e.g. VOD servers; Operations thereof
- H04N21/25—Management operations performed by the server for facilitating the content distribution or administrating data related to end-users or client devices, e.g. end-user or client device authentication, learning user preferences for recommending movies
- H04N21/251—Learning process for intelligent management, e.g. learning user preferences for recommending movies
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F16/00—Information retrieval; Database structures therefor; File system structures therefor
- G06F16/90—Details of database functions independent of the retrieved data types
- G06F16/95—Retrieval from the web
- G06F16/953—Querying, e.g. by the use of web search engines
- G06F16/9535—Search customisation based on user profiles and personalisation
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
- G06N3/049—Temporal neural networks, e.g. delay elements, oscillating neurons or pulsed inputs
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/08—Learning methods
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/40—Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
- H04N21/45—Management operations performed by the client for facilitating the reception of or the interaction with the content or administrating data related to the end-user or to the client device itself, e.g. learning user preferences for recommending movies, resolving scheduling conflicts
- H04N21/466—Learning process for intelligent management, e.g. learning user preferences for recommending movies
- H04N21/4667—Processing of monitored end-user data, e.g. trend analysis based on the log file of viewer selections
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/40—Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
- H04N21/45—Management operations performed by the client for facilitating the reception of or the interaction with the content or administrating data related to the end-user or to the client device itself, e.g. learning user preferences for recommending movies, resolving scheduling conflicts
- H04N21/466—Learning process for intelligent management, e.g. learning user preferences for recommending movies
- H04N21/4668—Learning process for intelligent management, e.g. learning user preferences for recommending movies for recommending content, e.g. movies
Landscapes
- Engineering & Computer Science (AREA)
- Databases & Information Systems (AREA)
- Theoretical Computer Science (AREA)
- Physics & Mathematics (AREA)
- Computing Systems (AREA)
- Multimedia (AREA)
- Signal Processing (AREA)
- General Physics & Mathematics (AREA)
- General Engineering & Computer Science (AREA)
- Data Mining & Analysis (AREA)
- Biophysics (AREA)
- Computational Linguistics (AREA)
- Biomedical Technology (AREA)
- Evolutionary Computation (AREA)
- General Health & Medical Sciences (AREA)
- Molecular Biology (AREA)
- Artificial Intelligence (AREA)
- Life Sciences & Earth Sciences (AREA)
- Mathematical Physics (AREA)
- Software Systems (AREA)
- Health & Medical Sciences (AREA)
- Two-Way Televisions, Distribution Of Moving Picture Or The Like (AREA)
Abstract
The invention provides a long-term and short-term interest modeling method for the IPTV field, which distinguishes a long-term click sequence Lt‑1And short-term click sequence StLong term click sequence Lt‑1And short-term click sequence StEmbedding mapping to obtain long-term click sequence embedded representationAnd short-term click sequence embedded representationFurther calculating the long-term preference of the family and the short-term interest U of the userSThen according to family long-term preference and user short-term interest USObtaining user mix preferences UhybirdAnd according to the user mixing preference UhybirdCalculating click rate Ru. According to the invention, more accurate interest and hobby prediction and push are realized through the operation.
Description
Technical Field
The invention belongs to the field of deep learning modeling, and particularly relates to a long-short term interest modeling method for the field of IPTV.
Background
In the era of information overload, recommendation systems have become very important for internet services and are also widely used in different fields, such as: e-commerce websites, video websites, and the like. The television, one of the most commonly used household appliances in daily life, has gradually developed towards the internet. Currently, a large amount of video content has been embedded in IPTV. Therefore, the IPTV field also needs to introduce a recommendation system to solve the problem of how to filter the content meeting the preference of the user. In the prior art, a method and system for modeling the interests of a television viewer is disclosed, for example, in patent application No. 201610485614.2.
However, the IPTV application and the video website in the internet mainly have the following two problems:
1) there is a lot of implicit feedback in IPTV applications, but the explicit feedback is very poor. This makes it impossible to tell whether the user dislikes or does not notice his or her non-interactive items;
2) more particularly, the user of IPTV is typically the entire home, not an individual. The preferences of each person in the family may be different, which greatly increases the difficulty of recommending tasks.
Disclosure of Invention
Aiming at the problems in the prior art, the invention provides a long-term and short-term interest modeling method for the IPTV field, which constructs the long-term preference of a family and the short-term interest of a current user according to the historical click records of the family group and the user by using a sequence modeling mode, thereby realizing more accurate recommendation.
The specific implementation content of the invention is as follows:
a long-short term interest modeling method for IPTV field divides the history click record into long-term click sequence Lt-1And short-term click sequence StProcessing the data and obtaining the long-term preference U of the family through self-attention calculationLUser short-term interest US。
In order to better implement the invention, the method further specifically comprises the following steps:
s1, obtaining a long-term click sequence L according to historical click recordst-1Short term click sequence St(ii) a Will long term click sequence Lt-1Short term click sequence StInputting the code into an embedded code layer to be coded, wherein the code is represented in a dense matrix form; the long-term click sequence obtained after coding is then embedded into the representationShort term click sequence embedded representationOutputting to a long-term and short-term modeling layer;
s2, embedding and representing long-term click sequences in long-term and short-term modeling layersShort term click sequence embedded representationSelf-attention calculation is carried out, and family long-term preference U is obtained through mapping respectivelyLUser short-term interest USAnd make the family prefer U for a long timeLUser short-term interest USInputting the long-short term preference fusion layer;
s3, long-term preference U of the family in the long-term and short-term preference fusion layerLUser short-term interest USPerforming self-attention decoding and mapping to obtain user mixed preference Uhybird;
S4, mixing preference U to usershybirdLinear mapping is carried out to obtain the click rate RuAnd (6) predicting.
In order to better implement the present invention, further, in the step s2, the short-term interest U of the user is obtainedSThe method comprises the following specific steps:
step SA. calculation of user short-term interests U in a self-attention mechanismSThe three inputs of Query, Key and Value are completely consistent and are short-term click sequencesColumn embedded representationEmbedding three-input short-term click sequences into a representationMapping into different spaces respectively through linear transformation to obtain embedded representation in corresponding spaceWherein Wq,Wk,WvHas the dimension ofTherefore, it isHas the dimension ofC is a fixed value;
then will beIs split into nhHead, obtained in the dimension ofTensor ofWherein d ═ C/nhThen transposing each tensor, the tensorFrom dimension ofIs transformed into
Step SB. divides each header after splittingAs an input of the dot product attention scaled, the following operations are performed: will be provided withPerforming matrix multiplication to obtain dimension ofThe dot product attention of (1); then reducing the dot product attention by square root times of dimensionality through scale transformation; then, performing softmax normalization processing on the accumulated attention; finally, the dot product is added with attentionMatrix multiplication outputs scaled dot product attention;
step SC. splices the scaled dot product attentiveness of each output, and then performs linear transformation to output short-term multi-head attentivenessShort term multi-headed attentionHas the dimension of
Step SD. directs short term multiple attentionObtaining user short-term interest U by input point type feedforward networkS。
To better implement the present invention, in step SD., before outputting the multi-headed attention to the point-type feedforward network, residual concatenation and layer normalization are performed, and then the short-term multi-headed attention is focusedInputting a point type feedforward network, and performing residual error connection and layer normalization on the output of the point type feedforward network to obtain the user short-term interest US。
In order to better implement the present invention, further, the step s2 obtains the long-term preference U of the familyLThe method comprises the following specific steps:
step Sa. for Long term click sequence Lt-1Encoded long-term click sequence embedded representation for each time stepThe long-term click sequence embedded representationHas the dimension ofEmbedding a representation for each long-term click sequenceShort-term multi-head attention is calculated as described aboveIs processed to obtain the long-term preference U for calculating the familyLLong-term multi-head attention ofThe long-term multi-head attentionHas the dimension of
Step Sb. combines the long-term click sequence Lt-1Long-term multi-head attention per step inIs compressed and then the average value of each column is calculated to obtain the dimension ofA row vector of (a); then the long-term click sequence Lt-1All the line vectors of the time step are spliced to finally obtain the dimension ofFamily long-term preference UL。
In order to better implement the present invention, further, the step s3. specifically includes: preference of family for long term ULUser short-term interest USCalculating multi-head attention through a common attention mechanism, performing residual connection and layer normalization on the processed output, inputting the result into a point type feedforward network for further processing, performing residual connection and layer normalization on the further processed result, and finally obtaining user mixed preference U after linear change and Sigmoid function processinghybirdThe user mix preferences UhybirdDimension of
In order to better implement the present invention, further, the specific operation of step s4 is: first mix the user preferences U by global average poolinghybirdReduce the dimension toThe click rate R is then obtained by linear mapping, i.e. matrix multiplicationuThe click rate RuHas the dimension of
To better implement the invention, further, the long-term click sequence Lt-1And short-term click sequence StDividing according to t time step: the t time step is the last time segment in the acquired click sequence to be processed, and all time steps before the t time step in the acquired click sequence to be processed are defined as a long-term click sequence Lt-1The long-term click sequence Lt-1=S1∪S2∪...∪St-1Defining t time step as a short-term click sequence StThe short-term click sequence St={i1,i2,...,im}。
To better implement the invention, further, the encoded long-term click sequence Lt-1The specific operation is as follows: will long term click sequence Lt-1Each time step S iniEmbedding mapping to obtain long-term click sequence embedded representationThe long-term click sequence embedded representationDimension ofWherein | SiI represents the number of media assets in the processed time step, and K represents the size of the embedding layer;
the encoded short-term click sequence StThe specific operation is as follows: short-term click sequence StEmbedding mapping to obtain short-term click sequence embedded representationThe short term click sequence embedded representationHas the dimension ofWherein | StAnd | represents the number of assets in the t time step, and K represents the size of the embedding layer.
In order to better implement the invention, further, the point feed-forward network is composed of two fully-coupled layers, and a ReLU activation function is set between the two fully-coupled layers.
Compared with the prior art, the invention has the following advantages and beneficial effects:
1) meanwhile, the hobbies of family groups and the interests of users in a short period are considered, so that the interest recommendation result is more accurate;
2) through the click sequence, the user can be better distinguished whether the user dislikes or does not notice the non-recommended videos;
3) self-attention computation is used, which refines the representation by matching individual sequences to themselves; unlike general attention, the mechanism of self-attention reduces reliance on external information and is better at capturing internal correlations of sequences or features.
Drawings
FIG. 1 is a flow chart of the present invention;
FIG. 2 is a model structure and operation diagram of the present invention;
FIG. 3 shows the long term preference U of a familyLUser short-term interest USCalculating a flow chart;
FIG. 4 is family long term preference ULUser short-term interest USA processing flow chart of multi-head attention in the calculation process;
FIG. 5 is a flow chart of the process of scaling dot product attention by ratio during a multi-head attention process;
FIG. 6 is a calculation of user mix preferences U for the decoding processhybirdIs described.
Detailed Description
In order to more clearly illustrate the technical solutions of the embodiments of the present invention, the technical solutions in the embodiments of the present invention will be clearly and completely described below with reference to the drawings in the embodiments of the present invention, and it should be understood that the described embodiments are only a part of the embodiments of the present invention, and not all embodiments, and therefore should not be considered as a limitation to the scope of protection. All other embodiments, which can be obtained by a person skilled in the art without any inventive step based on the embodiments of the present invention, are within the scope of the present invention.
Example 1:
a method for modeling long-short term interest in IPTV field, as shown in FIG. 1, comprises the following steps:
s1, obtaining a long-term click sequence L according to historical click recordst-1Short term click sequence St(ii) a The long-term click sequence Lt-1And short-term click sequence StDividing according to t time step: the t time step is the last time segment in the acquired click sequence to be processed, and in the acquired click sequence to be processed, the time step before the t time step is defined as a long-term click sequence Lt-1The long-term click sequence Lt-1=S1∪S2∪...∪St-1Defining t time step as a short-term click sequence StThe short-term click sequence St={i1,i2,...,im}; for long-term click sequence Lt-1Encoding is carried out, and the specific operation is as follows: will long term click sequence Lt-1Each time step S iniEmbedding mapping to obtain long-term click sequence embedded representationThe long-term click sequence embedded representationDimension ofWherein | SiI represents the number of media assets in the processed time step, and K represents the size of the embedding layer; for short-term click sequence StEncoding is carried out, and the specific operation is as follows: short-term click sequence StEmbedding mapping to obtain short-term click sequence embedded representationThe short term click sequence embedded representationHas the dimension ofWherein | StI represents the number of media assets in the time step of t, and K represents the size of the embedding layer; the long-term click sequence obtained after coding is then embedded into the representationShort term click sequence embedded representationOutputting to a long-term and short-term modeling layer;
s2, embedding and representing long-term click sequences in long-term and short-term modeling layersShort term click sequence embedded representationSelf-attention calculation is carried out, and family long-term preference U is obtained through mapping respectivelyLUser short-term interest USAnd make the family prefer U for a long timeLUser short-term interest USInputting the long-short term preference fusion layer;
s3, long-term preference U of the family in the long-term and short-term preference fusion layerLUser short-term interest USPerforming self-attention decoding and mapping to obtain user mixed preference Uhybird;
S4, mixing preference U to usershybirdLinear mapping is carried out to obtain the click rate RuAnd (6) predicting.
The working principle is as follows: through the division of the t time step, the preference of the family user before the t time step can be divided, the preference interest of the user using the television at present is the t time step, the preference interest and the preference interest are processed after the t time step is divided, and the final result can be more accurate. The specific division of the time step needs to be determined according to the scene, assuming that it is first time-dependentThe inter-step is divided in hours, so if the user clicks, the behavior of nearly one hour is considered as a short-term sequence and the earlier behavior is considered as a long-term sequence. Then coding the long-term and short-term sequence to obtain the family long-term preference ULAnd user short-term interests USThen using the family long-term preference ULDecoding user short-term interests USObtaining user mix preferences Uhybird. Finally, the user is mixed with the preference UhybirdLinear mapping is carried out to finally obtain the click rate R of the useruAnd (6) predicting. And in actual operation, in our recommendation system, there would be two phases, namely a recall phase and a sort phase, and the model would be used in the sort phase. First, in the recall phase, the recommendation system generates a longer candidate list, and then filters each candidate in the candidate list through the ranking model. Let us assume that the user clicks each candidate item and clicks, and calculate the click rate R according to the above-mentioned calculation procedureu. Finally the click rate RuAnd sequencing to obtain a final recommendation list.
Example 2:
the present invention is based on the above embodiment 1, as shown in FIG. 2, and further, as shown in FIG. 2, the present invention is to arrange the long-term click sequence Lt-1And short-term click sequence StAre respectively input into an embedded coding layer which maps the long-term click sequence L through embeddingt-1And short-term click sequence StEncoded as a long-term click sequence Lt-1Vector and short-term click sequence embedded representationThe long-term click sequence is then embedded into the representationAnd short-term click sequence embedded representationSending the data into a long-short term modeling layer, and carrying out long-term click sequence on the long-short term modeling layer through a Transformer encoderColumn embedded representationAnd short-term click sequence embedded representationCalculating the long-term preference U of the family by self attentionLAnd user short-term interests USNote that the representation is embedded for short-term click sequencesDirectly obtaining the short-term interest U of the user after being processed by a Transformer encoderSAnd for long-term click sequences the representation is embeddedAfter being processed by a Transformer encoder, the long-term preference U of the family can be obtained through global average poolingL(ii) a Because of the long-term click sequence Lt-1And short-term click sequence StDivided by t time steps, short-term click sequences StOne time step, and a long-term click sequence Lt-1For short-term click sequences StA plurality of time steps before the time step, so the long-term preference U of the family can be obtained after one global average poolingL. Family long-term preference U to be generatedLAnd user short-term interests USSending the information into the long and short term preference fusion layer, and carrying out attention calculation in a Transformer decoder of the fusion layer of the long and short term preference to obtain the user mixed preference UhybirdFinally, mix the user preferences UhybirdDimension reduction by global average pooling toThe click rate R is then obtained by linear mapping, i.e. matrix multiplicationuThe click rate RuHas the dimension of
Other parts of this embodiment are the same as those of embodiment 1, and thus are not described again.
Example 3:
in this embodiment, based on any one of the above embodiments 1-2, and as shown in fig. 3, 4, and 5, further, the ordinary attention mechanism has three inputs, which are Query (request), Key (primary Key), and Value (Value), and the calculation method is as follows:
and in the self-attention mechanism, for calculating the short-term interest U of the userSAre identical, all short term click sequence embedded representationsEmbedding three-input short-term click sequences into a representationRespectively by linear transformation Wq,Wk,WvMapping into different spacesWherein Wq,Wk,WvThe dimensions of (a) are respectively:therefore, it isThe dimensions of (a) are respectively:
then will beIs split into nhHead, obtainingDimension ofTensor ofWherein d ═ C/nhThen transposing each tensor, the tensorFrom m nhX d is converted into nh×m×d;
As shown in FIG. 3, the user's short-term interests U are calculated using self-attention in the long-and-short-term modeling layerSThe process of (2) is as follows: first embedding the short-term click sequence of the input into the representationTreatment as short-term multi-head attentionThen the obtained short-term attention of the headInput into a point feed-forward network, where attention is paid to the point feed-forward network and the short termThe output ends of the two-dimensional model are all provided with a residual error connection and are subjected to layer normalization processing, and the short-term interest U of the user can be obtained after the point type feedforward network, the residual error connection and the layer normalization processingSIt is used.
Short-term Bull attention with respect to FIG. 3The calculation method of (2) is, as shown in fig. 4, to make linear change to the inputted multi-head data, then to process it into scaled dot product attention, then to splice the multi-head scaled dot product attention, and finally to make linear transformationOutputting short-term multi-head attention
On the other hand, as for the processing method of dot product attention by scaling in fig. 4, as shown in fig. 5, the tensor divided into multiple tensorsAs an input of the dot product attention scaled, the following operations are performed: will be provided withPerforming matrix multiplication to obtain a dot product attention with dimension of m multiplied by m; then reducing the dot product attention by square root times of dimensionality through scale transformation; then, performing softmax normalization processing on the accumulated attention; finally, the dot product is added with attentionMatrix multiplication outputs scaled dot product attention;
likewise, the family long term preference U is calculated for the use of self-attention in the long and short term modeling layerLAnd the above-mentioned process for calculating the short-term interest U of the userSThe operation is approximately the same, and the long-term click sequence needs to be embedded with a representationWherein the short-term click sequence embedding representation E is performed for each time stepStOperation of the treatment to obtain long-term attentionThe difference is that global average pooling treatment is required, and the specific steps are as follows: long-term multi-head attention per time step will be obtainedCompressing the dimensionality, and then calculating the average value of each column to obtain a row vector; then click long termSequence Lt-1All the line vectors of the time step are spliced to finally obtain the long-term preference U of the familyL(ii) a The formula for the global average pooling is as follows:
wherein GAP represents Global Average Pooling (Global Average Pooling), ULPreference of family for long term UL,The output of attention for each step.
Other parts of this embodiment are the same as any of embodiments 1-2 described above, and thus are not described again.
Example 4:
this embodiment is different from the transform encoder in that the transform decoder uses a normal attention mechanism rather than a self-attention mechanism, based on any of the embodiments 1 to 3 described above, as shown in fig. 6. The user's short-term interest U is processed through similar calculation steps as a transform encoder, including a multi-head attention mechanism, residual connection and regularization and a point feed-forward networkSAnd inputting the family long-term interest preference to obtain the attention output of the transform decoder, wherein the specific user mixed preference UhybirdThe calculation formula of (2) is as follows:
Uhybird=Attention(UL,US,US)
After obtaining the user mixing preference UhybirdThen, the users are mixed with the preference U through global average poolinghybirdReducing dimension, and then obtaining click rate R through linear mapping, namely matrix multiplicationuThe resulting click rate RuHas the dimension ofThe specific calculation formula is as follows:
Ru=Linear(GAP(Uhybird))。
other parts of this embodiment are the same as any of embodiments 1 to 3, and thus are not described again.
The above description is only a preferred embodiment of the present invention, and is not intended to limit the present invention in any way, and all simple modifications and equivalent variations of the above embodiments according to the technical spirit of the present invention are included in the scope of the present invention.
Claims (4)
1. A long-term and short-term interest modeling method for the IPTV field is characterized in that a long-term click sequence L is extracted from a historical click recordt-1And short-term click sequence StProcessing the data and obtaining the long-term preference U of the family through self-attention calculationLUser short-term interest US(ii) a The method specifically comprises the following steps:
s1, obtaining a long-term click sequence L according to historical click recordst-1Short term click sequence St(ii) a Will long term click sequence Lt-1Short term click sequence StInputting the code into an embedded code layer to be coded, wherein the code is represented in a dense matrix form; the long-term click sequence obtained after coding is then embedded into the representationShort term click sequence embedded representationOutputting to a long-term and short-term modeling layer;
s2, embedding and representing long-term click sequences in long-term and short-term modeling layersShort term click sequence embedded representationSelf-attention calculation is carried out, and family long-term preference U is obtained through mapping respectivelyLUser short-term interest USAnd make the family prefer U for a long timeLUser short-term interest USInputting the long-short term preference fusion layer;
s3, long-term preference U of the family in the long-term and short-term preference fusion layerLUser short-term interest USPerforming self-attention decoding and mapping to obtain user mixed preference Uhybird;
S4, mixing preference U to usershybirdLinear mapping is carried out to obtain click rate R for predictionu;
In the step S2, the short-term interest U of the user is obtainedSThe method comprises the following specific steps:
step SA. calculation of user short-term interests U in a self-attention mechanismSThe three inputs of Query, Key and Value are completely consistent and are embedded expressions of short-term click sequencesEmbedding three-input short-term click sequences into a representationMapping into different spaces respectively through linear transformation to obtain embedded representation in corresponding spaceWherein the weight matrix W of the linear transformationq,Wk,WvHas the dimension ofTherefore, it isHas the dimension ofC is a fixed value;
then will beIs split into nhHead, obtained in the dimension ofTensor ofWherein d ═ C/nhThen transposing each tensor, the tensorFrom dimension ofIs transformed into
Step SB. divides each header after splittingThe matrix is used as an input of the dot product attention of scaling, and the following operations are carried out: will be provided withPerforming matrix multiplication to obtain dimension ofThe dot product attention of (1); then reducing the dot product attention by square root times of dimensionality through scale transformation; then, performing softmax normalization processing on the accumulated attention; finally, the dot product is added with attentionMatrix multiplication outputs scaled dot product attention;
step SC. splices the scaled dot product attentiveness of each output, and then performs linear transformation to output short-term multi-head attentivenessThe short-term multi-headed attentionHas the dimension of
Step SD. directs short term multiple attentionObtaining user short-term interest U by input point type feedforward networkS;
In said step SD., attention is paid to the short-term bullBefore inputting to the point type feedforward network, residual error connection and layer normalization are firstly carried out, and the residual error connection and layer normalization are also carried out on the output of the point type feedforward network, and finally the short-term interest U of the user is obtainedS;
The long-term preference U of the family is obtained in the step S2LThe method comprises the following specific steps:
step Sa. for Long term click sequence Lt-1Encoded long-term click sequence embedded representation for each time stepThe long-term click sequence embedded representationHas the dimension ofEmbedding a representation for each long-term click sequenceShort-term multi-head attention is calculated as described aboveIs processed to obtain the long-term preference U for calculating the familyLLong-term multi-head attention ofThe long-term multi-head attentionHas the dimension of
Step Sb. combines the long-term click sequence Lt-1Long-term multi-head attention per step inIs compressed and then the average value of each column is calculated to obtain the dimension ofA row vector of (a); then the long-term click sequence Lt-1All the line vectors of the time step are spliced to finally obtain the dimension ofFamily long-term preference UL;
The step S3 specifically comprises the following operations: preference of family for long term ULUser short-term interest USCalculating multi-head attention through a common attention mechanism, performing residual connection and layer normalization on the processed output, inputting the result into a point type feedforward network for further processing, performing residual connection and layer normalization on the further processed result, and finally obtaining user mixed preference U after linear change and Sigmoid function processinghybirdThe user mix preferences UhybirdDimension of
2. The method of claim 1, wherein the long-term click sequence L is a long-term short-term interest modeling method in IPTV fieldt-1And short-term click sequence StDividing according to t time step: the t time step is the last time segment in the acquired click sequence to be processed, and all time steps before the t time step in the acquired click sequence to be processed are defined as a long-term click sequence Lt-1The long-term click sequence Lt-1=S1∪S2∪...∪St-1Defining t time step as a short-term click sequence StThe short-term click sequence St={i1,i2,...,im}。
3. The method as claimed in claim 2, wherein the method for modeling the long-short term interest in IPTV field is characterized in thatThen, the code long-term click sequence Lt-1The specific operation is as follows: will long term click sequence Lt-1Each time step S iniEmbedding mapping to obtain long-term click sequence embedded representationThe long-term click sequence embedded representationDimension ofWherein | SiI represents the number of media assets in the processed time step, and K represents the size of the embedding layer;
the encoded short-term click sequence StThe specific operation is as follows: short-term click sequence StEmbedding mapping to obtain short-term click sequence embedded representationThe short term click sequence embedded representationHas the dimension ofWherein | StAnd | represents the number of assets in the t time step, and K represents the size of the embedding layer.
4. The method as claimed in claim 1, wherein the point feed-forward network comprises two fully-coupled layers, and a ReLU activation function is set between the two fully-coupled layers.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202010129277.XA CN111294619B (en) | 2020-02-28 | 2020-02-28 | Long-short term interest modeling method for IPTV field |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202010129277.XA CN111294619B (en) | 2020-02-28 | 2020-02-28 | Long-short term interest modeling method for IPTV field |
Publications (2)
Publication Number | Publication Date |
---|---|
CN111294619A CN111294619A (en) | 2020-06-16 |
CN111294619B true CN111294619B (en) | 2021-09-10 |
Family
ID=71029259
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN202010129277.XA Expired - Fee Related CN111294619B (en) | 2020-02-28 | 2020-02-28 | Long-short term interest modeling method for IPTV field |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN111294619B (en) |
Families Citing this family (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN114422859B (en) * | 2020-10-28 | 2024-01-30 | 贵州省广播电视信息网络股份有限公司 | Deep learning-based ordering recommendation system and method for cable television operators |
CN113344662A (en) * | 2021-05-31 | 2021-09-03 | 联想(北京)有限公司 | Product recommendation method, device and equipment |
Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN110008409A (en) * | 2019-04-12 | 2019-07-12 | 苏州市职业大学 | Based on the sequence of recommendation method, device and equipment from attention mechanism |
CN110489639A (en) * | 2019-07-15 | 2019-11-22 | 北京奇艺世纪科技有限公司 | A kind of content recommendation method and device |
CN110599280A (en) * | 2018-06-12 | 2019-12-20 | 阿里巴巴集团控股有限公司 | Commodity information preference model training and predicting method and device and electronic equipment |
CN110610168A (en) * | 2019-09-20 | 2019-12-24 | 合肥工业大学 | Electroencephalogram emotion recognition method based on attention mechanism |
CN110807156A (en) * | 2019-10-23 | 2020-02-18 | 山东师范大学 | Interest recommendation method and system based on user sequence click behaviors |
Family Cites Families (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN107995535B (en) * | 2017-11-28 | 2019-11-26 | 百度在线网络技术(北京)有限公司 | A kind of method, apparatus, equipment and computer storage medium showing video |
CN110162553A (en) * | 2019-05-21 | 2019-08-23 | 南京邮电大学 | Users' Interests Mining method based on attention-RNN |
CN110446112A (en) * | 2019-07-01 | 2019-11-12 | 南京邮电大学 | IPTV user experience prediction technique based on two-way LSTM-Attention |
CN110796313B (en) * | 2019-11-01 | 2022-05-31 | 北京理工大学 | Session recommendation method based on weighted graph volume and item attraction model |
-
2020
- 2020-02-28 CN CN202010129277.XA patent/CN111294619B/en not_active Expired - Fee Related
Patent Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN110599280A (en) * | 2018-06-12 | 2019-12-20 | 阿里巴巴集团控股有限公司 | Commodity information preference model training and predicting method and device and electronic equipment |
CN110008409A (en) * | 2019-04-12 | 2019-07-12 | 苏州市职业大学 | Based on the sequence of recommendation method, device and equipment from attention mechanism |
CN110489639A (en) * | 2019-07-15 | 2019-11-22 | 北京奇艺世纪科技有限公司 | A kind of content recommendation method and device |
CN110610168A (en) * | 2019-09-20 | 2019-12-24 | 合肥工业大学 | Electroencephalogram emotion recognition method based on attention mechanism |
CN110807156A (en) * | 2019-10-23 | 2020-02-18 | 山东师范大学 | Interest recommendation method and system based on user sequence click behaviors |
Non-Patent Citations (2)
Title |
---|
Adaptive User Modeling with Long and Short-Term Preferences for Personalized Recommendation;Zeping Yu;《Proceedings of the Twenty-Eighth International Joint Conference on Artificial Intelligence(IJCAI-19)》;20191231;全文 * |
基于神经网络的IPTV用户体验预测系统设计与实现;毛佳丽;《中国优秀硕士学位论文全文数据库 信息科技辑》;20200215;全文 * |
Also Published As
Publication number | Publication date |
---|---|
CN111294619A (en) | 2020-06-16 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
JP6523498B1 (en) | Learning device, learning method and learning program | |
Pfaff et al. | Neural network based intra prediction for video coding | |
CN111294619B (en) | Long-short term interest modeling method for IPTV field | |
CN111310045A (en) | Network-embedded movie recommendation method based on meta-path | |
CN107562787B (en) | POI (point of interest) encoding method and device, POI recommendation method and electronic equipment | |
Chang et al. | On the design fundamentals of diffusion models: A survey | |
CN113094587B (en) | Implicit recommendation method based on knowledge graph path | |
CN113255908A (en) | Method, neural network model and device for service prediction based on event sequence | |
CN114386513A (en) | Interactive grading prediction method and system integrating comment and grading | |
CN107506479B (en) | A kind of object recommendation method and apparatus | |
CN113033090A (en) | Push model training method, data push device and storage medium | |
CN114357201A (en) | Audio-visual recommendation method and system based on information perception | |
Sharifi et al. | A new algorithm for solving data sparsity problem based-on Non negative matrix factorization in recommender systems | |
He et al. | A time-context-based collaborative filtering algorithm | |
CN113918764A (en) | Film recommendation system based on cross modal fusion | |
CN117171440A (en) | News recommendation method and system based on news event and news style joint modeling | |
CN116680456A (en) | User preference prediction method based on graph neural network session recommendation system | |
CN117251622A (en) | Method, device, computer equipment and storage medium for recommending objects | |
CN111046280A (en) | Cross-domain recommendation method for application FM | |
CN114782995A (en) | Human interaction behavior detection method based on self-attention mechanism | |
CN111222722B (en) | Method, neural network model and device for business prediction for business object | |
CN111460302B (en) | Data processing method, device, electronic equipment and computer readable storage medium | |
CN112132345A (en) | Method and device for predicting user information of electric power company, electronic equipment and storage medium | |
Kumari et al. | Movie Recommendation System for Cold-Start Problem Using User's Demographic Data | |
Ahirwadkar et al. | Deepautoencf: A Denoising Autoencoder For Recommender Systems |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
GR01 | Patent grant | ||
GR01 | Patent grant | ||
CF01 | Termination of patent right due to non-payment of annual fee | ||
CF01 | Termination of patent right due to non-payment of annual fee |
Granted publication date: 20210910 |