CN108647226B - Hybrid recommendation method based on variational automatic encoder - Google Patents

Hybrid recommendation method based on variational automatic encoder Download PDF

Info

Publication number
CN108647226B
CN108647226B CN201810253803.6A CN201810253803A CN108647226B CN 108647226 B CN108647226 B CN 108647226B CN 201810253803 A CN201810253803 A CN 201810253803A CN 108647226 B CN108647226 B CN 108647226B
Authority
CN
China
Prior art keywords
user
vector
article
hidden
encoder
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN201810253803.6A
Other languages
Chinese (zh)
Other versions
CN108647226A (en
Inventor
张寅�
林建实
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Zhejiang University ZJU
Original Assignee
Zhejiang University ZJU
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Zhejiang University ZJU filed Critical Zhejiang University ZJU
Priority to CN201810253803.6A priority Critical patent/CN108647226B/en
Publication of CN108647226A publication Critical patent/CN108647226A/en
Application granted granted Critical
Publication of CN108647226B publication Critical patent/CN108647226B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06QINFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES; SYSTEMS OR METHODS SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES, NOT OTHERWISE PROVIDED FOR
    • G06Q30/00Commerce
    • G06Q30/06Buying, selling or leasing transactions
    • G06Q30/0601Electronic shopping [e-shopping]
    • G06Q30/0631Item recommendations
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks

Abstract

The invention discloses a hybrid recommendation method based on a variational automatic encoder. The method comprises the steps of modeling scoring characteristics and content characteristics of users and articles by using a variational automatic encoder, encoding sparse characteristics by a factorization machine, and automatically performing characteristic high-order combination; meanwhile, the multi-view data characteristics of the user and the article are fused into the framework of the variational automatic encoder so as to solve the problem of cold start; and through the variation inference analysis of the hidden vector codes of the user and the article, the method provides the interpretability for the automatic encoder to generate the hidden vector codes; by inputting the characteristics corresponding to the user and the articles, the preference values of the user to the candidate article set can be obtained, and the recommendation results are obtained by sorting according to the preference values. Compared with the traditional recommendation method, the method has better recommendation effect.

Description

Hybrid recommendation method based on variational automatic encoder
Technical Field
The invention relates to a computer recommendation system, in particular to a hybrid recommendation method based on an automatic encoder.
Background
In recent years, with the continuous development of networks and information technologies, the data volume, the generation speed and the complexity of online information are rapidly increased, and the personalized recommendation system has become an important technical means for extracting information from complex data and is widely applied in the industry.
Traditional collaborative filtering-based recommendation methods, particularly matrix factorization series methods, have proven to be very effective in the industry, and although implicit feedback data such as browsing, clicking and collecting are easier to collect than explicit feedback data such as movie scoring, merchandise evaluation, etc., the cold start problem and feature sparsity problem remain important factors that limit the performance of recommendation systems.
In recent years, deep learning has been developed in the fields of graphic images, natural language processing and the like, and has proved its excellent performance in feature processing, so that the application of deep learning to recommendation systems has become an important direction in this field. However, the existing models based on the deep neural network often process the content characteristics of users and articles, and the key of the recommendation system is to describe the interaction relationship between the users and the articles, on this point, the deep learning research is not directly applied, and more hidden vector codes generated are directly utilized to be input into a matrix decomposition frame.
Disclosure of Invention
Aiming at the blank and the defects of the prior art, the invention provides a hybrid recommendation method based on a variational automatic encoder. The technical scheme adopted by the invention is as follows:
the hybrid recommendation method based on the variational automatic encoder comprises the following steps:
(1) processing log data according to a specific application configuration environment to obtain interaction relation information of a user and an article, wherein the interaction relation information comprises implicit feedback and explicit feedback; performing feature processing for different information types: for implicit feedback data, the mark of interactive behavior is 1, otherwise, the mark is 0; recording the specific score value of the display feedback data, and then carrying out normalization processing on the characteristic value;
(2) multi-view information of a user and an article is collected respectively, wherein the multi-view information comprises user portrait information and article content information, and the cold start problem is solved;
(3) collecting recommendation feedback information which is not favored by a user except for the articles with historical behaviors, generating negative samples, and enabling the number of the positive and negative samples to be the same;
(4) constructing a model based on a variational automatic encoder hybrid recommendation method, performing gradient updating of variables in an alternate iteration mode, training the model, and storing final model parameters; for the objects and users with historical interactive behaviors, corresponding hidden vector codes are reserved;
(5) in the prediction stage, the user and the article which have the hidden vector coding are directly used as the input of a generalized matrix decomposition module in the model, and the preference value of the user to the specific article is calculated; for users and articles lacking the hidden vector codes, calculating the corresponding hidden vector codes through a trained model, and calculating preference values of the hidden vector codes;
(6) for a specific user, calculating preference values of the user to the articles in the candidate article set, and sequencing the preference values to obtain a recommended article list of the user;
and (3) regularly arranging logs and repeating the calculation models from (1) to (4) in the execution process of the method, and updating the hidden vector codes of the users and the articles.
Preferably, the step (3) includes: for each user, dividing the article into a positive sample and a negative sample according to the existing interactive behavior, and screening a part of negative samples for the article without the interactive record in a sampling mode.
Preferably, the model based on the variational automatic encoder hybrid recommendation method in the step (4) is composed of three modules in total, including a variational automatic encoder on the user side, a variational automatic encoder on the article side and a generalized matrix decomposition module, wherein the variational automatic encoder is divided into a decoder and an encoder; and (4) after receiving the user and article characteristic values and the corresponding positive and negative sample preference values obtained in the steps (2), (3) and (4), training the model.
Preferably, the gradient updating formula of the variables in the step (4) is as follows:
Figure BDA0001608481060000021
Figure BDA0001608481060000022
Figure BDA0001608481060000023
Figure BDA0001608481060000024
Figure BDA0001608481060000025
wherein phiuvuvAnd Ψ is an encoder parameter of the user-automated encoder, an encoder parameter of the commodity-automated encoder, a decoder parameter of the user-automated encoder, a decoder parameter of the commodity-automated encoder, and a parameter of the generalized matrix decomposition module, θ and Φ are an encoder module parameter and a decoder module parameter, ηuvΨRespectively, the rate of updating the parameters, Z, of the user-side autoencoder, the article-side autoencoder, and the generalized matrix decomposition moduleu,ZvRespectively, the hidden vector codes, X, being generated by an automatic user-side encoder and an automatic article-side encoderB,UBRespectively, a user multi-view feature of random gradient descent batch size B and a scoring feature of the user, YB,VBA multi-view characteristic of the item and a scoring characteristic of the item, respectively, of batch size B, U and V are scoring characteristics of the user and the item, respectively, fpooling(U),fpooling(V) user and object, respectivelyAnd outputting the scoring characteristics of the products after the pooling operation.
Preferably, the step (5) includes the steps of:
1) saving the model training parameter phi obtained after the step (4) is implementeduvuvAnd Ψ for developing a prediction;
2) for users and articles with interactive behaviors, directly reading the stored hidden vector codes; for unknown users and articles, the calculation of the hidden vector coding is carried out through an encoder part;
3) for the encoder part of the user, implicit vector encoding of user i
Figure BDA0001608481060000031
The calculation formula is as follows:
Figure BDA0001608481060000032
Figure BDA0001608481060000033
Figure BDA0001608481060000034
Figure BDA0001608481060000035
Figure BDA0001608481060000036
where g (-) is the activation function of each layer, ui,xiRespectively the scoring feature and the multi-view feature of user i,
Figure BDA0001608481060000037
and
Figure BDA0001608481060000038
respectively, user i generates a mean vector, a variance vector and an implicit vector code through a variational automatic coder,
Figure BDA0001608481060000039
is the output result vector of the k hidden layer when the hidden vector of the user is encoded,
Figure BDA00016084810600000310
the weight vector corresponding to the k hidden layer is calculated when the hidden vector of the user is coded, and is respectively used for processing the output of the hidden layer and the multi-view characteristic input,
Figure BDA00016084810600000311
calculating the k hidden layer bias term corresponding to the user when encoding the hidden vector, wherein k is 2,3 …, L is the number of hidden layers, and L is the number of hidden layers
Figure BDA00016084810600000312
Is output aiming at mean value vector when computing implicit vector coding of user
Figure BDA0001608481060000041
The weight term of (a) is,
Figure BDA0001608481060000042
is output aiming at mean value vector when computing implicit vector coding of user
Figure BDA0001608481060000043
The bias term of (a) is,
Figure BDA0001608481060000044
is output aiming at variance vector when computing implicit vector coding of user
Figure BDA0001608481060000045
The weight term of (a) is,
Figure BDA0001608481060000046
is output aiming at variance vector when computing implicit vector coding of user
Figure BDA0001608481060000047
The bias term of (d); ε is the number sampled in accordance with a normal distribution with a mean of 0 and a variance of 1;
4) for the encoder part of the article, implicit vector encoding of article i
Figure BDA0001608481060000048
The calculation formula is as follows:
Figure BDA0001608481060000049
Figure BDA00016084810600000410
Figure BDA00016084810600000411
Figure BDA00016084810600000412
Figure BDA00016084810600000413
where g (-) is the activation function of each layer, vi,yiRespectively the scoring feature and the multi-view feature of item i,
Figure BDA00016084810600000414
and
Figure BDA00016084810600000415
respectively, mean vector, variance vector and implicit vector codes generated by the item i through a variational automatic encoder,
Figure BDA00016084810600000416
is the output result vector of the k hidden layer when the hidden vector coding of the article is calculated,
Figure BDA00016084810600000417
the weight vector corresponding to the k hidden layer when calculating the hidden vector code of the article is respectively used for processing the output of the hidden layer and the multi-view characteristic input,
Figure BDA00016084810600000418
when calculating the hidden vector coding of the article, corresponding to the k hidden layer bias term, k is 2,3 …, L is the number of hidden layers, and L is the number of hidden layers
Figure BDA00016084810600000419
Is output relative to the mean vector when calculating the hidden vector code of the article
Figure BDA00016084810600000420
The weight term of (a) is,
Figure BDA00016084810600000421
is output relative to the mean vector when calculating the hidden vector code of the article
Figure BDA00016084810600000422
The bias term of (a) is,
Figure BDA00016084810600000423
is the square difference vector output when calculating the hidden vector code of the article
Figure BDA00016084810600000424
The weight term of (a) is,
Figure BDA00016084810600000425
is the square difference vector output when calculating the hidden vector code of the article
Figure BDA00016084810600000426
The weight term bias term of (1); ε is the mean of coincidence value of 0 and the variance of1 is sampled by a normal distribution;
5) calculating the scoring preference value of the user to the item, wherein the formula is as follows:
R=fΨ(Zu,Zv)
wherein ZuHidden vector coding for users, ZvFor implicit vector coding of articles, fΨ(. h) is a function fitted to a neural network architecture with Ψ as a parameter.
According to the method, the grading feature and the content feature of the user and the article are modeled by using the variational automatic encoder, meanwhile, the multi-view data feature of the user and the article is fused into the framework of the variational automatic encoder, and the preference value of the user to the candidate article set can be obtained and the recommendation result can be obtained through the variational inference analysis of the hidden vector encoding of the user and the article. Compared with the traditional recommendation method, the method has better recommendation effect.
Drawings
FIG. 1 is an overall model diagram of a hybrid recommendation method based on a variational auto-encoder;
fig. 2 is a diagram of a user-side variational autoencoder network architecture.
Detailed Description
In order to make the objects, technical solutions and advantages of the present invention more apparent, the present invention is further described in detail below with reference to the accompanying drawings.
The hybrid recommendation method based on the variational automatic encoder comprises the following steps:
(1) processing log data of the system according to a specific application configuration environment, and obtaining interaction relation information such as browsing, collection, clicking, comment and the like of a user and an article through construction of a data warehouse and cleaning of characteristic data, wherein the interaction relation information mainly comprises implicit feedback and explicit feedback; performing feature processing for different information types: for implicit feedback data, the mark of interactive behavior is 1, otherwise, the mark is 0; recording the specific score value of the display feedback data; then, normalizing the characteristic value;
(2) multi-view information of users and articles is collected respectively and managed in a data warehouse mode, so that the problem of cold start is solved: collecting user portrait information, such as user age, gender, school, specialty, past behavior records, etc.; collecting content information of the article, such as graphic features of pictures, features extracted by describing texts through a natural language processing method, click rate and collection rate of the article, and the like;
(3) collecting other recommendation feedback information of the user except for the articles with the historical behavior records, for example, the user presents preference to some articles, generating negative samples, and then using the negative samples to make the integral positive and negative sample numbers approximately same;
the method comprises the following steps: for each user, dividing the article into a positive sample and a negative sample according to the existing interactive behavior, and screening a part of negative samples for the article without the interactive record in a sampling mode.
(4) Constructing a model based on a variational automatic encoder hybrid recommendation method, performing gradient updating of variables in an alternate iteration mode, training the model, and storing final model parameters; for the objects and users with historical interactive behaviors, corresponding hidden vector codes are reserved;
in this step, the model construction process based on the variational automatic encoder hybrid recommendation method is as follows:
the model based on the variational automatic encoder hybrid recommendation method mainly comprises three parts, namely an automatic encoder framework on the left side and an automatic encoder framework on the right side and a Multi-Layer neural network framework in the middle, namely an MLP (Multi-Layer Perceptin) module in FIG. 1. Two automatic encoder frameworks are adopted to encode a user and an article respectively, and respectively generate hidden Vector representations (tension vectors) of the user and the article and a pooling layer result of a factorization machine as the input of a multilayer neural network. Multi-view feature x for simultaneous user and itemiAnd yjConcatenating into the input of each hidden layer of the Encoder (Encoder) module to enable the hidden vector representation of the user and the article to learn information from multiple view sources, when there is no corresponding scoring information for a new user or new articleAnd generating a hidden vector by the data of other views for score estimation, thereby relieving the cold start problem.
As shown in FIG. 2, taking the user-side encoder as an example, U is defined as user rating data, X is the characteristic of multi-view data of the user, and the characteristic concatenation is performed only on the encoder side, phiuAs parameter configuration on the encoder side, and ZuIs the intermediate layer of the generated implicit vector code, passed through the decoder ΘuSeparately reconstruct the scoring data
Figure BDA0001608481060000061
And multi-view data
Figure BDA0001608481060000062
The derivation is performed on the user side, the architecture on the article side is similar, and for the sake of symbolic representation, Θ is not distinguished in this sectionuAnd Θ. Process conditional probability P of reconstructing back to original inputs U and X from low-dimensional hidden vector Zθ(X, U | Z) denotes that θ is a parameter in the reconstruction process, and according to the maximum likelihood estimation, the goal is to maximize the likelihood probability P (X, U) ═ P (X, U; Z, θ), and the unknown implicit vector code Z and the reconstruction process θ are solved, so that the probability of reconstructing the original input X and U is maximized:
Figure BDA0001608481060000063
the posterior probability P (Z | X, U) of the hidden vector Z is not computable, and one way to vary the autoencoder is to introduce Qφ(Z | X, U) to approximate Pθ(Z|X,U)。
Specifically, on the right side of the above equation, + logQφ(Z|X,U)-logQφ(Z | X, U) is as follows:
Figure BDA0001608481060000071
both sides simultaneously pair Qφ(Z | X, U) is expected to:
Figure BDA0001608481060000072
the goal of maximum likelihood estimation is to make the likelihood probability P of a sampleθ(X, U) is maximum and due to Qφ(Z | X, U) is PθApproximate distribution of (Z | X, U), KL (Q)φ(Z|X,U)||Pθ(Z | X, U)) > 0, so:
Figure BDA0001608481060000073
then
Figure BDA0001608481060000074
Is the Lower Bound of likelihood probability, called the Variational Lower Bound (Variational Lower Bound).
Transformed by a bayesian formula:
Figure BDA0001608481060000075
maximizing the lower bound of likelihood probability requires approximating distribution Q using assumptionsφUnder (Z | X, U) conditions, the expected maximum of the probabilities of X and U can be generated, while the distribution used can be such that the assumed distribution approximates the a priori distribution of Z. Only the optimal solution theta and phi of the lower bound of the maximum likelihood probability is obtained, an automatic encoder can be designed, and P is calculatedθThe process of (X, U | Z) is represented as a generator, from a given Pθ(Z) in the case of prior probability distribution, X and U, i.e., the data points in the sample, are generated by the generator with the highest probability of occurrence when the reconstruction error for the original inputs X and U is the smallest.
Similarly, the lower bound of the optimization variation of the network architecture on the item side can be obtained, and for consistency of notation, the optimization objective on the user side is re-described, as follows:
Figure BDA0001608481060000076
Figure BDA0001608481060000077
consistent with the conventional matrix decomposition-based recommendation method, the distributions of p (z) and q (z | x, u) by assuming implicit vector coding are gaussian distributions. And then expanding to respectively obtain the optimization targets of the user and the article, namely the lower variation bound is as follows:
Figure BDA0001608481060000081
Figure BDA0001608481060000082
combining the optimization objective of the generalized matrix decomposition model, the final loss function is:
Figure BDA0001608481060000083
the model based on the hybrid recommendation method of the variational automatic encoder consists of three modules in total, including the variational automatic encoder at one side of a user, the variational automatic encoder at one side of an article and a generalized matrix decomposition module, wherein the variational automatic encoder is divided according to a decoder and an encoder, parameters of five parts are totally included, and an updated recursion formula is as follows:
Figure BDA0001608481060000084
Figure BDA0001608481060000085
Figure BDA0001608481060000086
Figure BDA0001608481060000087
Figure BDA0001608481060000088
wherein phiuvuvAnd Ψ is an encoder parameter of the user-automated encoder, an encoder parameter of the commodity-automated encoder, a decoder parameter of the user-automated encoder, a decoder parameter of the commodity-automated encoder, and a parameter of the generalized matrix decomposition module, θ and Φ are general names of an encoder module parameter and a decoder module parameter, ηuvΨRespectively, the rate of updating the parameters, Z, of the user-side autoencoder, the article-side autoencoder, and the generalized matrix decomposition moduleu,ZvRespectively, the hidden vector codes, X, being generated by an automatic user-side encoder and an automatic article-side encoderB,UBRespectively, a user multi-view feature of random gradient descent batch size B and a scoring feature of the user, YB,VBA multi-view characteristic of the item and a scoring characteristic of the item, respectively, of batch size B, U and V are scoring characteristics of the user and the item, respectively, fpooling(U),fpooling(V) is the output of the user and item scoring features after the pooling operation, respectively. And alternately and iteratively optimizing the parameters of the five parts by an optimization mode such as gradient descent and the like. Meanwhile, the final parameters of the five parts are reserved, and the hidden vector codes of the existing users and articles can be obtained through calculation and stored, so that the existing users and articles can be directly used as the input of the generalized matrix decomposition module when the existing users and articles are encountered subsequently, and the calculation is accelerated.
(5) In the prediction stage, the user and the article which have the hidden vector coding are directly used as the input of a generalized matrix decomposition module in the model, and the preference value of the user to the specific article is calculated; for users and articles lacking the hidden vector codes, calculating the corresponding hidden vector codes through a trained model, and calculating preference values of the hidden vector codes;
the method comprises the following substeps:
1) saving the model training parameter phi obtained after the step (4) is implementeduvuvAnd Ψ, develop a prediction;
2) for users and articles with interactive behaviors, directly reading the stored hidden vector codes; for unknown users and articles, calculating through an encoder part;
3) for the encoder part of the user, implicit vector encoding of user i
Figure BDA0001608481060000091
The calculation formula is as follows:
Figure BDA0001608481060000092
Figure BDA0001608481060000093
Figure BDA0001608481060000094
Figure BDA0001608481060000095
Figure BDA0001608481060000096
where g (-) is the activation function of each layer, ui,xiRespectively the scoring feature and the multi-view feature of user i,
Figure BDA0001608481060000097
and
Figure BDA0001608481060000098
respectively, user i generates a mean vector, a variance vector and an implicit vector code through a variational automatic coder,
Figure BDA0001608481060000099
is the output result vector of the k hidden layer when the hidden vector of the user is encoded,
Figure BDA0001608481060000101
the weight vector corresponding to the k hidden layer is calculated when the hidden vector of the user is coded, and is respectively used for processing the output of the hidden layer and the multi-view characteristic input,
Figure BDA0001608481060000102
calculating the k hidden layer bias term corresponding to the user when encoding the hidden vector, wherein k is 2,3 …, L is the number of hidden layers, and L is the number of hidden layers
Figure BDA0001608481060000103
Is output aiming at mean value vector when computing implicit vector coding of user
Figure BDA0001608481060000104
The weight term of (a) is,
Figure BDA0001608481060000105
is output aiming at mean value vector when computing implicit vector coding of user
Figure BDA0001608481060000106
The bias term of (a) is,
Figure BDA0001608481060000107
is output aiming at variance vector when computing implicit vector coding of user
Figure BDA0001608481060000108
The weight term of (a) is,
Figure BDA0001608481060000109
is to calculate the hour hand of the implicit vector coding of the userFor variance vector output
Figure BDA00016084810600001010
The bias term of (d); ε is the number sampled in accordance with a normal distribution with a mean of 0 and a variance of 1;
4) for the encoder part of the article, implicit vector encoding of article i
Figure BDA00016084810600001011
The calculation formula is as follows:
Figure BDA00016084810600001012
Figure BDA00016084810600001013
Figure BDA00016084810600001014
Figure BDA00016084810600001015
Figure BDA00016084810600001016
where g (-) is the activation function of each layer, vi,yiRespectively the scoring feature and the multi-view feature of item i,
Figure BDA00016084810600001017
and
Figure BDA00016084810600001018
respectively, mean vector, variance vector and implicit vector codes generated by the item i through a variational automatic encoder,
Figure BDA00016084810600001019
is the output result vector of the k hidden layer when the hidden vector coding of the article is calculated,
Figure BDA00016084810600001020
the weight vector corresponding to the k hidden layer when calculating the hidden vector code of the article is respectively used for processing the output of the hidden layer and the multi-view characteristic input,
Figure BDA00016084810600001021
when calculating the hidden vector coding of the article, corresponding to the k hidden layer bias term, k is 2,3 …, L is the number of hidden layers, and L is the number of hidden layers
Figure BDA00016084810600001022
Is output relative to the mean vector when calculating the hidden vector code of the article
Figure BDA00016084810600001023
The weight term of (a) is,
Figure BDA00016084810600001024
is output relative to the mean vector when calculating the hidden vector code of the article
Figure BDA00016084810600001025
The bias term of (a) is,
Figure BDA00016084810600001026
is the square difference vector output when calculating the hidden vector code of the article
Figure BDA00016084810600001027
The weight term of (a) is,
Figure BDA00016084810600001028
is the square difference vector output when calculating the hidden vector code of the article
Figure BDA00016084810600001029
The weight term bias term of (1); ε is the value sampled in accordance with a normal distribution with a mean of 0 and a variance of 1;
5) Calculating the scoring preference value of the user to the item, wherein the formula is as follows:
R=fΨ(Zu,Zv)
wherein ZuHidden vector coding for users, ZvFor implicit vector coding of articles, fΨ(. h) is a function fitted to the neural network architecture with Ψ as a parameter, which can be chosen to correspond to the form desired.
(6) For a specific user, calculating preference values of the user to the articles in the candidate article set, and sequencing the preference values to obtain a recommended article list of the user;
in the whole recommendation method execution process, the system logs are generated continuously, so that the logs need to be regularly sorted, the calculation models from (1) to (4) need to be repeated, and the hidden vector codes of the users and the articles need to be updated.

Claims (5)

1. A hybrid recommendation method based on a variational automatic encoder is characterized by comprising the following steps:
(1) processing log data according to a specific application configuration environment to obtain interaction relation information of a user and an article, wherein the interaction relation information comprises implicit feedback and explicit feedback; performing feature processing for different information types: for implicit feedback data, the mark of interactive behavior is 1, otherwise, the mark is 0; recording the specific score value of the display feedback data, and then carrying out normalization processing on the characteristic value;
(2) collecting multi-view information of a user and an article respectively, wherein the multi-view information comprises user portrait information and article content information;
(3) collecting recommendation feedback information which is not favored by a user except for the articles with historical behaviors, generating negative samples, and enabling the number of the positive and negative samples to be the same;
(4) constructing a model based on a variational automatic encoder hybrid recommendation method, performing gradient updating of variables in an alternate iteration mode, training the model, and storing final model parameters; for the objects and users with historical interactive behaviors, corresponding hidden vector codes are reserved;
(5) in the prediction stage, the user and the article which have the hidden vector coding are directly used as the input of a generalized matrix decomposition module in the model, and the preference value of the user to the specific article is calculated; for users and articles lacking the hidden vector codes, calculating the corresponding hidden vector codes through a trained model, and calculating preference values of the hidden vector codes;
(6) for a specific user, calculating preference values of the user to the articles in the candidate article set, and sequencing the preference values to obtain a recommended article list of the user;
and (3) regularly arranging logs and repeating the calculation models from (1) to (4) in the execution process of the method, and updating the hidden vector codes of the users and the articles.
2. The variational automatic encoder-based hybrid recommendation method according to claim 1, wherein said step (3) comprises: for each user, dividing the article into a positive sample and a negative sample according to the existing interactive behavior, and screening a part of negative samples for the article without the interactive record in a sampling mode.
3. The variational automatic encoder-based hybrid recommendation method according to claim 1, wherein the variational automatic encoder-based hybrid recommendation method model in step (4) is composed of a total of three modules, including a user-side variational automatic encoder, an article-side variational automatic encoder and a generalized matrix decomposition module, wherein the variational automatic encoder is divided into a decoder and an encoder; and (4) after receiving the user and article characteristic values and the corresponding positive and negative sample preference values obtained in the steps (2), (3) and (4), training the model.
4. The method for recommending a mixture based on a variational automatic encoder according to claim 1, wherein said step (4) comprises the following gradient update formula of variables:
Figure FDA0003247342360000021
Figure FDA0003247342360000022
Figure FDA0003247342360000023
Figure FDA0003247342360000024
Figure FDA0003247342360000025
wherein phiu,Φv,Θu,ΘvAnd Ψ is an encoder parameter of the user-automated encoder, an encoder parameter of the commodity-automated encoder, a decoder parameter of the user-automated encoder, a decoder parameter of the commodity-automated encoder, and a parameter of the generalized matrix decomposition module, θ and Φ are an encoder module parameter and a decoder module parameter, ηu,ηv,ηΨRespectively, the rate of updating the parameters, Z, of the user-side autoencoder, the article-side autoencoder, and the generalized matrix decomposition moduleu,ZvRespectively, the hidden vector codes, X, being generated by an automatic user-side encoder and an automatic article-side encoderB,UBRespectively, a user multi-view feature of random gradient descent batch size B and a scoring feature of the user, YB,VBA multi-view characteristic of the item and a scoring characteristic of the item, respectively, of batch size B, U and V are scoring characteristics of the user and the item, respectively, fpooling(U),fpooling(V) output of the user and item scoring features after pooling operations, respectively;
Figure FDA0003247342360000026
a parameter gradient representing a batch size B;
Figure FDA0003247342360000027
representing a global parameter gradient.
5. The variational automatic encoder-based hybrid recommendation method according to claim 1, characterized in that said step (5) comprises the steps of:
1) saving the model training parameter phi obtained after the step (4) is implementedu,Φv,Θu,ΘvAnd Ψ for developing a prediction;
2) for users and articles with interactive behaviors, directly reading the stored hidden vector codes; for unknown users and articles, the calculation of the hidden vector coding is carried out through an encoder part;
3) for the encoder part of the user, implicit vector encoding of user i
Figure FDA0003247342360000031
The calculation formula is as follows:
Figure FDA0003247342360000032
Figure FDA0003247342360000033
Figure FDA0003247342360000034
Figure FDA0003247342360000035
Figure FDA0003247342360000036
where g (-) is the activation function of each layer, ui,xiRespectively the scoring feature and the multi-view feature of user i,
Figure FDA0003247342360000037
and
Figure FDA0003247342360000038
respectively, user i generates a mean vector, a variance vector and an implicit vector code through a variational automatic coder,
Figure FDA0003247342360000039
is the output result vector of the k-th hidden layer during the calculation of the hidden vector encoding of the user, Wk (en),Vk (en)The weight vector corresponding to the k hidden layer is calculated when the hidden vector of the user is coded, and is respectively used for processing the output of the hidden layer and the multi-view characteristic input,
Figure FDA00032473423600000310
calculating a hidden vector coding of a user corresponding to a k hidden layer bias term, wherein k is 2,3, L is the number of hidden layers, and W is the number of hidden layersL (μ),VL (μ)Is output aiming at mean value vector when computing implicit vector coding of user
Figure FDA00032473423600000311
The weight term of (a) is,
Figure FDA00032473423600000312
is output aiming at mean value vector when computing implicit vector coding of user
Figure FDA00032473423600000313
Bias term of (1), WL (σ),VL (σ)Is output aiming at variance vector when computing implicit vector coding of user
Figure FDA00032473423600000314
The weight term of (a) is,
Figure FDA00032473423600000315
is output aiming at variance vector when computing implicit vector coding of user
Figure FDA00032473423600000316
The bias term of (d); ε is the number sampled in accordance with a normal distribution with a mean of 0 and a variance of 1;
4) for the encoder part of the article, implicit vector encoding of article i
Figure FDA00032473423600000317
The calculation formula is as follows:
Figure FDA00032473423600000318
Figure FDA00032473423600000319
Figure FDA00032473423600000320
Figure FDA00032473423600000322
Figure FDA00032473423600000321
wherein g (. cndot.) is the excitation of each layerLive function, vi,yiRespectively the scoring feature and the multi-view feature of item i,
Figure FDA0003247342360000041
and
Figure FDA0003247342360000042
respectively, mean vector, variance vector and implicit vector codes generated by the item i through a variational automatic encoder,
Figure FDA0003247342360000043
is the output result vector of the k hidden layer when the hidden vector coding of the article is calculated,
Figure FDA0003247342360000044
the weight vector corresponding to the k hidden layer when calculating the hidden vector code of the article is respectively used for processing the output of the hidden layer and the multi-view characteristic input,
Figure FDA0003247342360000045
when the hidden vector coding of the object is calculated, corresponding to a k hidden layer bias term, k is 2,3, L is the number of hidden layers, and L is the number of hidden layers
Figure FDA0003247342360000046
Is output relative to the mean vector when calculating the hidden vector code of the article
Figure FDA0003247342360000047
The weight term of (a) is,
Figure FDA0003247342360000048
is output relative to the mean vector when calculating the hidden vector code of the article
Figure FDA0003247342360000049
The bias term of (a) is,
Figure FDA00032473423600000410
is the square difference vector output when calculating the hidden vector code of the article
Figure FDA00032473423600000411
The weight term of (a) is,
Figure FDA00032473423600000412
is the square difference vector output when calculating the hidden vector code of the article
Figure FDA00032473423600000413
The weight term bias term of (1); ε is the number sampled in accordance with a normal distribution with a mean of 0 and a variance of 1;
5) calculating the scoring preference value of the user to the item, wherein the formula is as follows:
R=fΨ(Zu,Zv)
wherein ZuHidden vector coding for users, ZvFor implicit vector coding of articles, fΨ(. h) is a function fitted to a neural network architecture with Ψ as a parameter.
CN201810253803.6A 2018-03-26 2018-03-26 Hybrid recommendation method based on variational automatic encoder Active CN108647226B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201810253803.6A CN108647226B (en) 2018-03-26 2018-03-26 Hybrid recommendation method based on variational automatic encoder

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201810253803.6A CN108647226B (en) 2018-03-26 2018-03-26 Hybrid recommendation method based on variational automatic encoder

Publications (2)

Publication Number Publication Date
CN108647226A CN108647226A (en) 2018-10-12
CN108647226B true CN108647226B (en) 2021-11-02

Family

ID=63744507

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201810253803.6A Active CN108647226B (en) 2018-03-26 2018-03-26 Hybrid recommendation method based on variational automatic encoder

Country Status (1)

Country Link
CN (1) CN108647226B (en)

Families Citing this family (10)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109543066B (en) * 2018-10-31 2021-04-23 北京达佳互联信息技术有限公司 Video recommendation method and device and computer-readable storage medium
CN109408729B (en) * 2018-12-05 2022-02-08 广州市百果园信息技术有限公司 Recommended material determination method and device, storage medium and computer equipment
CN110659411B (en) * 2019-08-21 2022-03-11 桂林电子科技大学 Personalized recommendation method based on neural attention self-encoder
CN110765353B (en) * 2019-10-16 2022-03-08 腾讯科技(深圳)有限公司 Processing method and device of project recommendation model, computer equipment and storage medium
US11915121B2 (en) 2019-11-04 2024-02-27 International Business Machines Corporation Simulator-assisted training for interpretable generative models
CN111709231B (en) * 2020-04-30 2022-11-18 昆明理工大学 Class case recommendation method based on self-attention variational self-coding
CN112231582B (en) * 2020-11-10 2023-11-21 南京大学 Website recommendation method and equipment based on variation self-coding data fusion
CN112188487B (en) * 2020-12-01 2021-03-12 索信达(北京)数据技术有限公司 Method and system for improving user authentication accuracy
CN113536116B (en) * 2021-06-29 2023-11-28 中国海洋大学 Cross-domain recommendation method based on double-stream sliced wasserstein self-encoder
CN115809374B (en) * 2023-02-13 2023-04-18 四川大学 Method, system, device and storage medium for correcting mainstream deviation of recommendation system

Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN107424016A (en) * 2017-08-10 2017-12-01 安徽大学 The real time bid method and its system that a kind of online wanted advertisement is recommended
CN107533683A (en) * 2015-04-28 2018-01-02 微软技术许可有限责任公司 Relevant group suggestion

Family Cites Families (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US11836746B2 (en) * 2014-12-02 2023-12-05 Fair Isaac Corporation Auto-encoder enhanced self-diagnostic components for model monitoring

Patent Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN107533683A (en) * 2015-04-28 2018-01-02 微软技术许可有限责任公司 Relevant group suggestion
CN107424016A (en) * 2017-08-10 2017-12-01 安徽大学 The real time bid method and its system that a kind of online wanted advertisement is recommended

Non-Patent Citations (7)

* Cited by examiner, † Cited by third party
Title
Augmented Variational Autoencoders for Collaborative Filtering with Auxiliary Information;Wonsung Lee等;《CIKM "17: Proceedings of the 2017 ACM on Conference on Information and Knowledge Management》;20171110;1139-1148 *
Cold-start, warm-start and everything in between: An autoencoder based approach to recommendation;Angshul Majumdar等;《2017 International Joint Conference on Neural Networks (IJCNN)》;20170703;3656-3663 *
Collaborative Variational Autoencoder for Recommender Systems;Xiaopeng Li等;《Proceedings of the 23rd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining》;20170817;305-314 *
基于深度学习的推荐系统研究综述;黄立威等;《计算机学报》;20180305(第07期);191-219 *
基于自动编码器的协同过滤推荐算法;俞晨光等;《微型电脑应用》;20151120(第11期);18-23 *
基于降噪自编码器网络与词向量的信息推荐方法;郭喻栋等;《计算机工程》;20171215(第12期);179-184 *
栈式降噪自编码器的标签协同过滤推荐算法;霍欢等;《小型微型计算机系统》;20180115(第01期);9-13 *

Also Published As

Publication number Publication date
CN108647226A (en) 2018-10-12

Similar Documents

Publication Publication Date Title
CN108647226B (en) Hybrid recommendation method based on variational automatic encoder
CN111274398B (en) Method and system for analyzing comment emotion of aspect-level user product
CN111275521B (en) Commodity recommendation method based on user comment and satisfaction level embedding
CN112487143B (en) Public opinion big data analysis-based multi-label text classification method
CN111127146B (en) Information recommendation method and system based on convolutional neural network and noise reduction self-encoder
CN112667818B (en) GCN and multi-granularity attention fused user comment sentiment analysis method and system
US20040243548A1 (en) Dependency network based model (or pattern)
CN111222332A (en) Commodity recommendation method combining attention network and user emotion
CN112328900A (en) Deep learning recommendation method integrating scoring matrix and comment text
CN109190109B (en) Method and device for generating comment abstract by fusing user information
CN107357899B (en) Short text sentiment analysis method based on sum-product network depth automatic encoder
CN111582506A (en) Multi-label learning method based on global and local label relation
CN113127737A (en) Personalized search method and search system integrating attention mechanism
CN112529071A (en) Text classification method, system, computer equipment and storage medium
CN114692605A (en) Keyword generation method and device fusing syntactic structure information
CN116077942A (en) Method for realizing interactive content recommendation
CN110019796A (en) A kind of user version information analysis method and device
CN117334271A (en) Method for generating molecules based on specified attributes
Liu et al. AutoDC: Automated data-centric processing
CN116070025A (en) Interpretable recommendation method based on joint score prediction and reason generation
CN114662652A (en) Expert recommendation method based on multi-mode information learning
CN115129807A (en) Fine-grained classification method and system for social media topic comments based on self-attention
CN114610871A (en) Information system modeling analysis method based on artificial intelligence algorithm
CN114357284A (en) Crowdsourcing task personalized recommendation method and system based on deep learning
CN111882441A (en) User prediction interpretation Treeshap method based on financial product recommendation scene

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant