CN110458627B - Commodity sequence personalized recommendation method for dynamic preference of user - Google Patents

Commodity sequence personalized recommendation method for dynamic preference of user Download PDF

Info

Publication number
CN110458627B
CN110458627B CN201910763463.6A CN201910763463A CN110458627B CN 110458627 B CN110458627 B CN 110458627B CN 201910763463 A CN201910763463 A CN 201910763463A CN 110458627 B CN110458627 B CN 110458627B
Authority
CN
China
Prior art keywords
commodity
user
layer
comment
vector
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN201910763463.6A
Other languages
Chinese (zh)
Other versions
CN110458627A (en
Inventor
黄震华
汤庸
刘海
李丁丁
蔡立群
廖晓鹏
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Guangdong Zhixin Information Technology Co ltd
South China Normal University
Original Assignee
Guangdong Zhixin Information Technology Co ltd
South China Normal University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Guangdong Zhixin Information Technology Co ltd, South China Normal University filed Critical Guangdong Zhixin Information Technology Co ltd
Priority to CN201910763463.6A priority Critical patent/CN110458627B/en
Publication of CN110458627A publication Critical patent/CN110458627A/en
Application granted granted Critical
Publication of CN110458627B publication Critical patent/CN110458627B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/30Information retrieval; Database structures therefor; File system structures therefor of unstructured textual data
    • G06F16/35Clustering; Classification
    • G06F16/353Clustering; Classification into predefined classes
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/90Details of database functions independent of the retrieved data types
    • G06F16/95Retrieval from the web
    • G06F16/953Querying, e.g. by the use of web search engines
    • G06F16/9535Search customisation based on user profiles and personalisation
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/21Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
    • G06F18/214Generating training patterns; Bootstrap methods, e.g. bagging or boosting
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/24Classification techniques
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06QINFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES; SYSTEMS OR METHODS SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES, NOT OTHERWISE PROVIDED FOR
    • G06Q30/00Commerce
    • G06Q30/02Marketing; Price estimation or determination; Fundraising
    • G06Q30/0241Advertisements
    • G06Q30/0251Targeted advertisements
    • G06Q30/0255Targeted advertisements based on user history
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06QINFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES; SYSTEMS OR METHODS SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES, NOT OTHERWISE PROVIDED FOR
    • G06Q30/00Commerce
    • G06Q30/02Marketing; Price estimation or determination; Fundraising
    • G06Q30/0241Advertisements
    • G06Q30/0251Targeted advertisements
    • G06Q30/0269Targeted advertisements based on user profile or attribute
    • G06Q30/0271Personalized advertisement
    • YGENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
    • Y02TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
    • Y02PCLIMATE CHANGE MITIGATION TECHNOLOGIES IN THE PRODUCTION OR PROCESSING OF GOODS
    • Y02P90/00Enabling technologies with a potential contribution to greenhouse gas [GHG] emissions mitigation
    • Y02P90/30Computing systems specially adapted for manufacturing

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Business, Economics & Management (AREA)
  • Data Mining & Analysis (AREA)
  • Physics & Mathematics (AREA)
  • Accounting & Taxation (AREA)
  • Development Economics (AREA)
  • General Engineering & Computer Science (AREA)
  • Finance (AREA)
  • Strategic Management (AREA)
  • Databases & Information Systems (AREA)
  • Game Theory and Decision Science (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • Marketing (AREA)
  • Economics (AREA)
  • Entrepreneurship & Innovation (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Artificial Intelligence (AREA)
  • General Business, Economics & Management (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Evolutionary Biology (AREA)
  • Evolutionary Computation (AREA)
  • Information Retrieval, Db Structures And Fs Structures Therefor (AREA)
  • Management, Administration, Business Operations System, And Electronic Commerce (AREA)

Abstract

The invention discloses a personalized commodity sequence recommendation method for dynamic preference of users, which comprises the steps of extracting commodity sequences under the condition of similar scores of the same user to construct commodity scoring vectors, extracting the user sequences with similar scores of the same commodity to obtain the user scoring vectors, combining personal information of the user, basic attribute information of the commodity, comments of the user and the commodity and commodity pictures, respectively realizing feature extraction of the user and the commodity based on multi-task learning, taking the feature vectors of the user and the historical commodity sequences thereof as input, realizing generation of the commodity sequences through a training encoder-decoder, and accurately learning recommendation of the optimal commodity sequences by combining search strategies. Based on multi-mode user-commodity data, the method and the device for optimizing the commodity sequence based on the multi-mode user-commodity data are used for highly extracting and fusing the user characteristics and the commodity characteristics, realizing personalized recommendation of the commodity sequence oriented to user preference and improving user experience.

Description

Commodity sequence personalized recommendation method for dynamic preference of user
Technical Field
The invention relates to the field of intelligent business, in particular to a commodity sequence personalized recommendation method for dynamic preference of users.
Background
With the advent of the big data age, a large number of users are generating a large amount of information every moment, and these information items are very huge, and how to effectively extract effective features from huge data to make recommendation is a problem to be solved urgently. Information preferences are different between different users in the recommendation field, but there is a certain correlation between similar users and goods. How to put forward a high-efficiency reasonable personalized recommendation method under the background of big data is a very worthy research hotspot.
The existing recommendation algorithm often uses a relation matrix between a user and a commodity and combines other characteristics between the user and the commodity to recommend, and although a certain reference can be provided for a recommendation result, the recommendation systems have the following disadvantages: the method has low processing efficiency on large-data-volume users and article information, dynamic historical preferences of the users and evaluation of historical commodities by the users are easily ignored, and the conventional recommendation results cannot be fed back well; secondly, the traditional recommendation system has the advantages of simple model, single user characteristic, certain limitation in evaluation by adopting a probability statistical method, long recommendation time and larger error; the traditional recommendation system is not combined with the picture features of the commodities, has single task, cannot qualitatively analyze the historical comments of the user and the pictures of the commodities, and cannot effectively evaluate the recommendation features; finally, the sequence information between the commodity sequences with the same user scores and the corresponding commodity sequences is ignored in the traditional recommendation algorithm, and the recommendation results are not reasonably ordered and fused, so that the overall recommendation effect is poor.
Disclosure of Invention
The invention mainly aims to provide a commodity sequence personalized recommendation method oriented to dynamic preference of users, and aims to overcome the problems.
In order to achieve the above purpose, the invention provides a commodity sequence personalized recommendation method for dynamic preference of users, which comprises the following steps:
s10, obtaining personal information of a user, commodity attribute information, commodity pictures, scores of the same user corresponding to different commodities, scores of different users of the same commodity, comment texts of the same user on the commodities and comment texts of the same commodity from the scoring information of the user on the commodities;
s20, constructing a user scoring vector generator according to the scoring of different users received by the same commodity, and generating a user scoring vector; constructing a commodity grading vector generator according to the grading corresponding to different commodities by the same user, and generating commodity grading vectors;
s30, inputting comment texts of the same user on all commodities into a user scoring vector generator to generate user comment vectors, and carrying out mean value clustering on the user comment vectors to generate user comment categories; inputting comment texts of users born by the same commodity into a commodity grading vector generator to generate commodity comment vectors, and carrying out mean value clustering on the commodity comment vectors to generate commodity comment categories; extracting picture feature vectors of the commodity by adopting a depth convolution network, and carrying out mean value clustering on the picture feature vectors to generate commodity categories corresponding to the picture;
S40, inputting the personal information of the user and the scoring vector of the user into a user feature extraction network for joint training to obtain a user ID and a user comment prediction type; inputting commodity attribute information and commodity grading vectors into a commodity feature extraction network for joint training to obtain commodity IDs, commodity comment prediction categories and commodity picture prediction categories;
s50, inputting the feature vectors of the user feature vectors and the feature vectors of the historical commodity sequences into a commodity sequence recommendation network, training by a training encoder-decoder in the commodity sequence recommendation network to generate user feature vectors corresponding to each user ID, taking the commodity feature vectors corresponding to each commodity ID as personalized recommended commodity sequences, and learning by combining a search strategy to obtain the recommendation of the optimal commodity sequences.
Preferably, the S30 includes:
the sentence embedded type end 2Vec tool takes Chinese Wikipedia data as a training corpus, extracts each comment text in the training corpus into one-dimensional comment vectors by adopting the sentence embedded type end 2Vec tool, performs t-means clustering on all the comment vectors to obtain N types of user comment categories, extracts one-dimensional commodity comment vectors from the comment text under the same commodity by adopting the sentence embedded type end 2Vec tool, and performs t-means clustering on all the commodity comment vectors to obtain M commodity comment categories; and extracting feature vectors of the commodity pictures by adopting a deep convolution network VGG16, and t-means clustering the comment vectors to obtain I commodity picture categories.
Preferably, the scoring information of the users used by the user scoring vector generator is divided into 5 groups, and the construction method of the user scoring vector generator comprises the following steps:
grouping into 5 groups according to the user scoring information under the same commodity, wherein the users with the same score are in the same group to obtain a user composition sequence with the same score of the same commodity<u 1 ,u 2 ,u 3 ,…u x >X is 50, which means that 50 users are selected for each group, and the sequence is selected<u 1 ,u 2 ,u 3 ,...,ur,…u x >As training sample, use user u r As input, divide user u in the output prediction sequence r Scoring vectors of other users, such that other user idsIf the probability product of (a) is maximum, if the total number of users is w=8000, the loss function y of the user scoring vector generator 1 The method comprises the following steps:
Figure BDA0002171116580000031
obtaining the scoring direction length of each user to be 200;
the construction method of the commodity grading vector generator comprises the following steps:
dividing the same user with the same scoring commodity into a group, and obtaining the commodity sequence with the same scoring from each group<i 1 ,i 2 ,i 3 ,…,i g ,...,i j >J is 20, which means that 20 commodities are selected in each group, and the commodity sequences are selected<i 1 ,i 2 ,i 3 ,…,i g ,...,i j >As training sample, commodity i g As input, divide i in the output prediction sequence g If the total number of commodities R=120000, obtaining the loss function y of the commodity grading vector generator by using a maximum likelihood estimation method 2 The method comprises the following steps:
Figure BDA0002171116580000032
the score vector length for each commodity was found to be 200.
Preferably, the user feature extraction network in S40 includes:
a first convolution layer adopts 2 convolution kernels with the size of 3*1, and the sliding step length is 1;
the first user pooling layer window size is 2*1;
the second user convolution layer adopts 4 convolution layers with the size of 3*1 and the convolution kernel step length of 2; the third user convolution layer adopts 5 convolution kernels with the size of 3*1, and the sliding step length is 1;
the fourth user convolution layer adopts 10 convolution kernels with the size of 5*1, and the step length of the convolution kernels is 2;
the pooling layer window size of the second user pooling layer is 2*1;
the fifth user convolution layer adopts 6 convolution kernels with the size of 3*1, and the step length of the convolution kernels is 2;
the window size of the third user pooling layer is 4*1;
the number of neurons in the user full connectivity layer is 200,
the user randomly deactivates the Dropout layer,
the user personal information and the user scoring vector are convolved through a first user convolution layer, the convolved vector is pooled through a first user pooling layer, then enters a third user convolution layer through batch standardization after being convolved through a second user convolution layer, enters a fourth user convolution layer after being convolved through batch standardization again, enters a user random inactivation Dropout layer after being output by the fourth user convolution layer, enters a second user pooling layer, then enters a user random inactivation Dropout layer for being output, enters a fifth user layer convolution layer, and then enters a user full connection layer after being pooled through the third user pooling layer, and the user characteristic vector is output.
Preferably, the user feature vector is accessed to a user index maximum function softmax layer after being output by a full connection layer, the softmax layer has 8000 nodes, 8000 users in the system are represented, the formed user set is U, the user feature vector is accessed to a comment index maximum function softmax layer after being output by the full connection layer, the softmax layer has 30 nodes, 30 user comment categories are represented, the formed user comment category set is T, the user index maximum function softmax layer and the comment index maximum function softmax layer both adopt a softmax multi-classification loss function,
let the current user be u r Obtaining user u through a user index maximum function softmax layer r Prediction of (2)
Probability P (u) r ):
Figure BDA0002171116580000041
Wherein sx (u) r ) Is u r Input of corresponding node in user index maximum function softmax layerEntering a value;
the comment category of the current user is z r Obtaining a user comment category z through a comment index maximum function softmax layer r Is P (z) r ):
Figure BDA0002171116580000042
Wherein sx (z) r ) Is z r Input values of corresponding nodes in the comment index maximum function softmax layer;
the total loss function of the user feature extraction network is:
Figure BDA0002171116580000043
wherein eta is L 1 The value of the regulating coefficient is between 0.7 and 0.8.
Preferably, the personal information of the user includes at least the following 9 attributes: the user age, the user gender, the user occupation, the region where the user is located, the average score, the purchase amount, the average browsing time, the collection amount and the average purchase amount are respectively represented by independent heat vectors, all the attributes are spliced into a long vector, and the long vector is reduced to a vector with the length of 50 through five layers of self-encoder tools; the commodity attribute information at least comprises the following 9 attributes: daily order volume, sales volume, collection volume, online time, use season, average life, average score, price interval.
Preferably, the commodity feature extraction network in S40 includes:
the first commodity convolution layer adopts 5 convolution kernels with the size of 3*1 and the step length of 2;
a second commodity convolution layer which adopts 73 convolution kernels with the size of 3*1, and the sliding step length is 3;
a first commodity pooling layer having a window size of 2*1;
a third commodity convolution layer which adopts 10 convolution kernels with the size of 5*1 and the step length of 3;
a second commodity pooling layer having a window size of 4*1;
a fourth commodity convolution layer which adopts 5 convolution kernels with the step length of 2 and the size of 3*1;
a third commodity pooling layer with a window size of 2*1;
a commodity random inactivation layer;
a fifth commodity convolution layer, which adopts 6 convolution kernels with the size of 3*1 and the step length of 3;
a commodity full-connection layer, the number of the neurons of which is 200,
the commodity attribute information and the commodity grading vector are input into a first convolution layer of a commodity feature extraction network and then are connected into a second convolution layer, the second convolution layer is input into a first pooling layer after output and is input into a third convolution layer after input into a commodity random inactivation layer after input into the second pooling layer after output, the second pooling layer is input into a fourth convolution layer after output, the fourth convolution layer is input into the third pooling layer after output, the third pooling layer is input into the commodity random inactivation layer after output into a fifth convolution layer, the commodity full-connection layer is input after output into the fifth convolution layer, and finally the commodity feature vector is output.
Preferably, the commodity feature vector is output from the commodity full-connection layer and then is connected to a commodity first-index maximum-function softmax layer, the softmax layer has 120000 nodes, 120000 commodities in the system are represented, the formed user set is V, the commodity feature vector is output from the commodity full-connection layer and then is connected to a commodity second-index maximum-function softmax layer, the softmax layer has 15 nodes, 15 commodity types are represented, and the formed commodity type set is A; the commodity feature vector is output from the commodity full-connection layer and then is connected to a commodity third index maximum function softmax layer, the softmax layer is provided with 50 nodes, 50 commodity picture categories are represented, and a formed commodity picture category set is I; the commodity feature vector is output from the commodity full-connection layer and then is connected into a commodity fourth index maximum function softmax layer, the softmax layer is provided with 30 nodes, 30 commodity comment categories are represented, the formed commodity comment category set is M, the commodity first index maximum function softmax layer, the commodity second index maximum function softmax layer, the commodity third index maximum function softmax layer and the commodity fourth index maximum function softmax layer all adopt softmax multi-classification loss functions,
Let the current commodity be i x Obtaining commodity i through a commodity first index maximum function softmax layer x Is P (i) x ):
Figure BDA0002171116580000061
Wherein sx (i) x ) Is i x Input values for corresponding nodes in the first softmax layer;
let the current commodity category be a x Obtaining commodity category a through commodity second index maximum function softmax layer x Is P (a) x ):
Figure BDA0002171116580000062
Wherein sx (a) x ) Is a as x Input values for corresponding nodes in the second softmax layer;
let the current commodity picture category be v x Obtaining commodity picture category v through commodity second index maximum function softmax layer x Is P (v) x ):
Figure BDA0002171116580000063
Wherein sx (v) x ) V is x Input values of corresponding nodes in the third softmax layer;
let the current commodity comment category be m x Obtaining commodity comment category m through commodity second index maximum function softmax layer x Is P (m) x ):
Figure BDA0002171116580000064
Wherein sx (m) x ) Is m x The input values of the corresponding nodes in the fourth softmax layer,
the total loss function of the commodity feature extraction network is:
Figure BDA0002171116580000065
wherein lambda is 1 ~λ 4 Is L 2 And lambda is the adjustment coefficient of (1) 1234 =1,λ 1 The value is between 0.1 and 0.3, lambda 2 The value is between 0.2 and 0.5, lambda 3 The value is between 0.2 and 0.4, lambda 4 The value is between 0.2 and 0.3.
Preferably, the step S50 specifically includes:
acquiring g commodities currently purchased by a user, and firstly using the previous g in the g commodities 1 The commodity sequence composed of the individual commodities is used as the input of the encoder and passes g 1 The gating cyclic GRU unit obtains the background vector of the encoder, then the background vector is spliced with the user characteristic vector and is used as the input of a decoder, and the decoder is composed of g-g 1 The gate control circulation GRU units are sequentially linked to form output nodes, and the output nodes respectively correspond to the rear g-g in g commodities 1 The number of the neural networks of each gating cycle GRU unit in the commodity sequence, the encoder and the decoder is set to be 1000, and the length of the background vector generated by the encoder is 200;
for the decoder, assume that the 20 product sequences output are < i 1 ,i 2 ,...,i 20 >, let s k (1 is less than or equal to k is less than or equal to 20) is commodity i k Corresponding decoder hidden variables s k =GRU(i k-1 ,c,s k-1 ) Then commodity i k Output probability p (i) k ) The representation is:
p(i k |i 1 ,i 2 ,...,i k-1 ,c)=p(i k-1 ,s k ,c),
the joint probability of the maximized output sequence of the kth node of the decoder is obtained as follows:
Figure BDA0002171116580000071
the loss function of the decoder output sequence is: y is 3 =-log P(i 1 ,i 2 ,...,i k ),
Adopting a Beam Search strategy Beam Search, setting the Beam Size beam_Size to be 2, taking an evaluation index as a return, calculating a return value generated by each commodity sequence on each decoding node, selecting the commodity sequence with the largest return value as a current recommended sequence, and calculating the return value through a normalized damage accumulated gain index NDCG, a Recall index Recall and an average precision average value index MAP to obtain the commodity sequence:
reward(i 1 ,i 2 ,i k )=0.4×NDCG(i 1 ,i 2 ,i k )+0.4×Recall(i 1 ,i 2 ,i k )+0.2×MAP(i 1 ,i 2 ,i k ),
The total loss function of the encoder is thus expressed as:
L 3 =-log P(i 1 ,i 2 ,...,i k )-reward(i 1 ,i 2 ,...,i k ),
back propagation is carried out according to the error of the recommended sequence, and the weight parameter of the encoder-decoder is adjusted;
the feature vector of each user and the feature vector of each commodity in the commodity sequence recommendation training are stored in a database, the user feature vector of each user is stored in a guide table of the database by taking the user id as a first column and the corresponding user feature vector as a second column, the commodity feature vector of each commodity is stored in an entry table of the database by taking the commodity id as a first column and the corresponding commodity feature vector as a second column.
Preferably, after S50, the method further includes:
s60, acquiring g currently purchased by the user according to the purchase log file 1 The commodity sequence of each commodity is obtained by inquiring the databaseFeature vectors of the commodity are input to an encoder in time sequence to generate background vectors; inquiring a database to obtain a characteristic vector of a current user, splicing a background vector and the characteristic vector of the current user, inputting the spliced background vector and the characteristic vector of the current user into a decoder, and generating g-g 1 And recommending the commodity sequence formed by the commodities to the current user as an optimal commodity sequence.
Compared with the prior art, the invention has the beneficial effects that:
1. The multi-mode characteristics of the user and the commodity are introduced, the comment, the score and the personal information of the user are effectively used, and meanwhile, the score, the comment, the picture and the attribute information of the commodity are fully utilized, so that the recommendation result is more accurate and reliable.
2. Based on deep learning and multitask learning technology, the invention respectively fuses the user and commodity features through the deep convolutional neural network, and adopts a multitask learning mode to jointly train a plurality of classification tasks, thereby being capable of deeply fusing the user features and commodity features instead of simply splicing.
3. The invention adopts an efficient sequence-to-sequence generation model, is based on the user history commodity sequence, and is integrated with the user characteristic vector, so that personalized recommendation can be realized.
4. The invention has the advantages of clear structure, reasonable logic, lower coupling degree between modules, easy realization and deployment, and capability of being rapidly expanded into a distributed and parallelized development environment, being beneficial to expansion, test maintenance and improving user experience.
Drawings
In order to more clearly illustrate the embodiments of the present invention or the technical solutions in the prior art, the drawings that are required in the embodiments or the description of the prior art will be briefly described, and it is obvious that the drawings in the following description are only some embodiments of the present invention, and other drawings may be obtained according to the structures shown in these drawings without inventive effort for a person skilled in the art.
Figure 1 is a logic structure diagram of the commodity sequence personalized recommendation offline training of the invention,
the achievement of the objects, functional features and advantages of the present invention will be further described with reference to the accompanying drawings, in conjunction with the embodiments.
Detailed Description
The following description of the embodiments of the present invention will be made clearly and fully with reference to the accompanying drawings, in which it is evident that the embodiments described are only some, but not all embodiments of the invention. All other embodiments, which can be made by those skilled in the art based on the embodiments of the invention without making any inventive effort, are intended to be within the scope of the invention.
It should be noted that, if directional indications (such as up, down, left, right, front, and rear … …) are included in the embodiments of the present invention, the directional indications are merely used to explain the relative positional relationship, movement conditions, etc. between the components in a specific posture (as shown in the drawings), and if the specific posture is changed, the directional indications are correspondingly changed.
In addition, if there is a description of "first", "second", etc. in the embodiments of the present invention, the description of "first", "second", etc. is for descriptive purposes only and is not to be construed as indicating or implying a relative importance or implicitly indicating the number of technical features indicated. Thus, a feature defining "a first" or "a second" may explicitly or implicitly include at least one such feature. In addition, the technical solutions of the embodiments may be combined with each other, but it is necessary to base that the technical solutions can be realized by those skilled in the art, and when the technical solutions are contradictory or cannot be realized, the combination of the technical solutions should be considered to be absent and not within the scope of protection claimed in the present invention.
As shown in fig. 1, the method for personalized recommendation of commodity sequences for dynamic preference of users provided by the invention comprises the following steps:
s10, obtaining personal information of a user, commodity attribute information, commodity pictures, scores of the same user corresponding to different commodities, scores of different users of the same commodity, comment texts of the same user on the commodities and comment texts of the same commodity from the scoring information of the user on the commodities;
s20, constructing a user scoring vector generator according to the scoring of different users received by the same commodity, and generating a user scoring vector; constructing a commodity grading vector generator according to the grading corresponding to different commodities by the same user, and generating commodity grading vectors;
s30, inputting comment texts of the same user on all commodities into a user scoring vector generator to generate user comment vectors, and carrying out mean value clustering on the user comment vectors to generate user comment categories; inputting comment texts of users born by the same commodity into a commodity grading vector generator to generate commodity comment vectors, and carrying out mean value clustering on the commodity comment vectors to generate commodity comment categories; extracting picture feature vectors of the commodity by adopting a depth convolution network, and carrying out mean value clustering on the picture feature vectors to generate commodity categories corresponding to the picture;
S40, inputting the personal information of the user and the scoring vector of the user into a user feature extraction network for joint training to obtain a user ID and a user comment prediction type; inputting commodity attribute information and commodity grading vectors into a commodity feature extraction network for joint training to obtain commodity IDs, commodity comment prediction categories and commodity picture prediction categories;
s50, inputting the feature vectors of the user feature vectors and the feature vectors of the historical commodity sequences into a commodity sequence recommendation network, training by a training encoder-decoder in the commodity sequence recommendation network to generate user feature vectors corresponding to each user ID, taking the commodity feature vectors corresponding to each commodity ID as personalized recommended commodity sequences, and learning by combining a search strategy to obtain the recommendation of the optimal commodity sequences.
The multi-mode characteristics of the user and the commodity are introduced, the comment, the score and the personal information of the user are effectively used, and meanwhile, the score, the comment, the picture and the attribute information of the commodity are fully utilized, so that the recommendation result is more accurate and reliable. Based on deep learning and multitask learning technology, the invention respectively fuses the user and commodity features through the deep convolutional neural network, and adopts a multitask learning mode to jointly train a plurality of classification tasks, thereby being capable of deeply fusing the user features and commodity features instead of simply splicing. The invention adopts an efficient sequence-to-sequence generation model, is based on the user history commodity sequence, and is integrated with the user characteristic vector, so that personalized recommendation can be realized. The invention has the advantages of clear structure, reasonable logic, lower coupling degree between modules, easy realization and deployment, and capability of being rapidly expanded into a distributed and parallelized development environment, and is beneficial to expansion and test maintenance.
Preferably, the S30 includes:
the sentence embedded type end 2Vec tool takes Chinese Wikipedia data as a training corpus, extracts each comment text in the training corpus into one-dimensional comment vectors by adopting the sentence embedded type end 2Vec tool, performs t-means clustering on all the comment vectors to obtain N types of user comment categories, extracts one-dimensional commodity comment vectors from the comment text under the same commodity by adopting the sentence embedded type end 2Vec tool, and performs t-means clustering on all the commodity comment vectors to obtain M commodity comment categories; and extracting feature vectors of the commodity pictures by adopting a deep convolution network VGG16, and t-means clustering the comment vectors to obtain I commodity picture categories.
Preferably, the scoring information of the users used by the user scoring vector generator is divided into 5 groups, and the construction method of the user scoring vector generator comprises the following steps:
grouping into 5 groups according to the user scoring information under the same commodity, wherein the users with the same score are in the same group to obtain a user composition sequence with the same score of the same commodity<u 1 ,u 2 ,u 3 ,…u x >X is 50, which means that 50 users are selected for each group, and the sequence is selected<u 1 ,u 2 ,u 3 ,...,u r ,…u x >As training sample, use user u r As input, divide user u in the output prediction sequence r Other users 'scoring vectors are excluded so that the probability product of other user ids is maximized, and if the total number of users w=8000, the user scoring vector generator's loss function y 1 The method comprises the following steps:
Figure BDA0002171116580000101
obtaining the scoring direction length of each user to be 200;
the construction method of the commodity grading vector generator comprises the following steps:
dividing the same user with the same scoring commodity into a group, and obtaining the commodity sequence with the same scoring from each group<i 1 ,i 2 ,i 3 ,…,i g ,...,i j >J is 20, which means that 20 commodities are selected in each group, and the commodity sequences are selected<i 1 ,i 2 ,i 3 ,…,i g ,...,i j >As training sample, commodity i g As input, divide i in the output prediction sequence g If the total number of commodities R=120000, obtaining the loss function y of the commodity grading vector generator by using a maximum likelihood estimation method 2 The method comprises the following steps:
Figure BDA0002171116580000111
the score vector length for each commodity was found to be 200.
Preferably, the user feature extraction network in S40 includes:
a first convolution layer adopts 2 convolution kernels with the size of 3*1, and the sliding step length is 1;
the first user pooling layer window size is 2*1;
a second user convolution layer adopts 4 convolution kernels with the size of 3*1 and the convolution kernel step length of 2
A layer; the third user convolution layer adopts 5 convolution kernels with the size of 3*1, and the sliding step length is 1;
The fourth user convolution layer adopts 10 convolution kernels with the size of 5*1, and the step length of the convolution kernels is 2;
the pooling layer window size of the second user pooling layer is 2*1;
the fifth user convolution layer adopts 6 convolution kernels with the size of 3*1, and the step length of the convolution kernels is 2;
the window size of the third user pooling layer is 4*1;
the number of neurons in the user full connectivity layer is 200,
the user randomly deactivates the Dropout layer,
the user personal information and the user scoring vector are convolved through a first user convolution layer, the convolved vector is pooled through a first user pooling layer, then enters a third user convolution layer through batch standardization after being convolved through a second user convolution layer, enters a fourth user convolution layer after being convolved through batch standardization again, enters a user random inactivation Dropout layer after being output by the fourth user convolution layer, enters a second user pooling layer, then enters a user random inactivation Dropout layer for being output, enters a fifth user layer convolution layer, and then enters a user full connection layer after being pooled through the third user pooling layer, and the user characteristic vector is output.
Preferably, the user feature vector is accessed to a user index maximum function softmax layer after being output by a full connection layer, the softmax layer has 8000 nodes, 8000 users in the system are represented, the formed user set is U, the user feature vector is accessed to a comment index maximum function softmax layer after being output by the full connection layer, the softmax layer has 30 nodes, 30 user comment categories are represented, the formed user comment category set is T, the user index maximum function softmax layer and the comment index maximum function softmax layer both adopt a softmax multi-classification loss function,
Let the current user be u r Obtaining user u through a user index maximum function softmax layer r Prediction of (2)
Probability P (u) r ):
Figure BDA0002171116580000112
Wherein sx (u) r ) Is u r Input values of corresponding nodes in a user index maximum function softmax layer;
the comment category of the current user is z r By comment on the maximum function of the indexThe softmax layer gets user comment category z r Is P (z) r ):
Figure BDA0002171116580000121
Wherein sx (z) r ) Is z r Input values of corresponding nodes in the comment index maximum function softmax layer;
the total loss function of the user feature extraction network is:
Figure BDA0002171116580000122
wherein eta is L 1 The value of the regulating coefficient is between 0.7 and 0.8.
Preferably, the personal information of the user includes at least the following 9 attributes: the user age, the user gender, the user occupation, the region where the user is located, the average score, the purchase amount, the average browsing time, the collection amount and the average purchase amount are respectively represented by independent heat vectors, all the attributes are spliced into a long vector, and the long vector is reduced to a vector with the length of 50 through five layers of self-encoder tools; the commodity attribute information at least comprises the following 9 attributes: daily order volume, sales volume, collection volume, online time, use season, average life, average score, price interval.
Preferably, the commodity feature extraction network in S40 includes:
the first commodity convolution layer adopts 5 convolution kernels with the size of 3*1 and the step length of 2;
a second commodity convolution layer which adopts 73 convolution kernels with the size of 3*1, and the sliding step length is 3;
a first commodity pooling layer having a window size of 2*1;
a third commodity convolution layer which adopts 10 convolution kernels with the size of 5*1 and the step length of 3;
a second commodity pooling layer having a window size of 4*1;
a fourth commodity convolution layer which adopts 5 convolution kernels with the step length of 2 and the size of 3*1;
a third commodity pooling layer with a window size of 2*1;
a commodity random inactivation layer;
a fifth commodity convolution layer, which adopts 6 convolution kernels with the size of 3*1 and the step length of 3;
a commodity full-connection layer, the number of the neurons of which is 200,
the commodity attribute information and the commodity grading vector are input into a first convolution layer of a commodity feature extraction network and then are connected into a second convolution layer, the second convolution layer is input into a first pooling layer after output and is input into a third convolution layer after input into a commodity random inactivation layer after input into the second pooling layer after output, the second pooling layer is input into a fourth convolution layer after output, the fourth convolution layer is input into the third pooling layer after output, the third pooling layer is input into the commodity random inactivation layer after output into a fifth convolution layer, the commodity full-connection layer is input after output into the fifth convolution layer, and finally the commodity feature vector is output.
Preferably, the commodity feature vector is output from the commodity full-connection layer and then is connected to a commodity first-index maximum-function softmax layer, the softmax layer has 120000 nodes, 120000 commodities in the system are represented, the formed user set is V, the commodity feature vector is output from the commodity full-connection layer and then is connected to a commodity second-index maximum-function softmax layer, the softmax layer has 15 nodes, 15 commodity types are represented, and the formed commodity type set is A; the commodity feature vector is output from the commodity full-connection layer and then is connected to a commodity third index maximum function softmax layer, the softmax layer is provided with 50 nodes, 50 commodity picture categories are represented, and a formed commodity picture category set is I; the commodity feature vector is output from the commodity full-connection layer and then is connected into a commodity fourth index maximum function softmax layer, the softmax layer is provided with 30 nodes, 30 commodity comment categories are represented, the formed commodity comment category set is M, the commodity first index maximum function softmax layer, the commodity second index maximum function softmax layer, the commodity third index maximum function softmax layer and the commodity fourth index maximum function softmax layer all adopt softmax multi-classification loss functions,
Let the current commodity be i x Obtaining commodity i through a commodity first index maximum function softmax layer x Is P (i) x ):
Figure BDA0002171116580000131
Wherein sx (i) x ) Is i x Input values for corresponding nodes in the first softmax layer;
let the current commodity category be a x Obtaining commodity category a through commodity second index maximum function softmax layer x Is P (a) x ):
Figure BDA0002171116580000132
Wherein sx (a) x ) Is a as x Input values for corresponding nodes in the second softmax layer;
let the current commodity picture category be v x Obtaining commodity picture category v through commodity second index maximum function softmax layer x Is P (v) x ):
Figure BDA0002171116580000133
Wherein sx (v) x ) V is x Input values of corresponding nodes in the third softmax layer;
let the current commodity comment category be m x Obtaining commodity comment category m through commodity second index maximum function softmax layer x Is P (m) x ):
Figure BDA0002171116580000141
Wherein sx (m) x ) Is m x Input of corresponding node in fourth softmax layerThe value of the sum of the values,
the total loss function of the commodity feature extraction network is:
Figure BDA0002171116580000142
wherein lambda is 1 ~λ 4 Is L 2 And lambda is the adjustment coefficient of (1) 1234 =1,λ 1 The value is between 0.1 and 0.3, lambda 2 The value is between 0.2 and 0.5, lambda 3 The value is between 0.2 and 0.4, lambda 4 The value is between 0.2 and 0.3.
Preferably, the step S50 specifically includes:
acquiring g commodities currently purchased by a user, and firstly using the previous g in the g commodities 1 The commodity sequence composed of the individual commodities is used as the input of the encoder and passes g 1 The gating cyclic GRU unit obtains the background vector of the encoder, then the background vector is spliced with the user characteristic vector and is used as the input of a decoder, and the decoder is composed of g-g 1 The gate control circulation GRU units are sequentially linked to form output nodes, and the output nodes respectively correspond to the rear g-g in g commodities 1 The number of the neural networks of each gating cycle GRU unit in the commodity sequence, the encoder and the decoder is set to be 1000, and the length of the background vector generated by the encoder is 200;
for the decoder, assume that the 20 product sequences output are < i 1 ,i 2 ,...,i 20 >, let s k (1 is less than or equal to k is less than or equal to 20) is commodity i k Corresponding decoder hidden variables s k =GRU(i k-1 ,c,s k-1 ) Then commodity i k Output probability p (i) k ) The representation is:
p(i k |i 1 ,i 2 ,...,i k-1 ,c)=p(i k-1 ,s k ,c),
the joint probability of the maximized output sequence of the kth node of the decoder is obtained as follows:
Figure BDA0002171116580000143
the loss function of the decoder output sequence is: y is 3 =-log P(i 1 ,i 2 ,...,i k ),
Adopting a Beam Search strategy Beam Search, setting the Beam Size beam_Size to be 2, taking an evaluation index as a return, calculating a return value generated by each commodity sequence on each decoding node, selecting the commodity sequence with the largest return value as a current recommended sequence, and calculating the return value through a normalized damage accumulated gain index NDCG, a Recall index Recall and an average precision average value index MAP to obtain the commodity sequence:
reward(i 1 ,i 2 ,i k )=0.4×NDCG(i 1 ,i 2 ,i k )+0.4×Recall(i 1 ,i 2 ,i k )+0.2×MAP(i 1 ,i 2 ,i k ),
The total loss function of the encoder is thus expressed as:
L 3 =-log P(i 1 ,i 2 ,...,i k )-reward(i 1 ,i 2 ,...,i k ),
back propagation is carried out according to the error of the recommended sequence, and the weight parameter of the encoder-decoder is adjusted;
the feature vector of each user and the feature vector of each commodity in the commodity sequence recommendation training are stored in a database, the user feature vector of each user is stored in a guide table of the database by taking the user id as a first column and the corresponding user feature vector as a second column, the commodity feature vector of each commodity is stored in an entry table of the database by taking the commodity id as a first column and the corresponding commodity feature vector as a second column.
Preferably, after S50, the method further includes:
s60, acquiring g currently purchased by the user according to the purchase log file 1 The commodity sequence of each commodity is inquired in a database to obtain the characteristic vector of each commodity in the commodity sequence, and the characteristic vectors are input to an encoder in time sequence to generate a background vector; querying a database to obtain a current userIs input into a decoder after the background vector and the characteristic vector of the current user are spliced, and g-g is generated 1 And recommending the commodity sequence formed by the commodities to the current user as an optimal commodity sequence.
Actual operation example of the present invention:
The invention is composed of a first commodity sequence personalized recommendation offline training module and a second commodity sequence personalized recommendation online application module.
In the first stage (multi-mode big data preprocessing) of the first module, after the original data of the recommendation system are cleaned and tidied, effective 8000 pieces of user data and 120000 pieces of commodity data are selected, commodity categories are 15, effective comments are about 250000 pieces, effective scores are 3000000 pieces, wherein the scores are divided into 1-5 grades, the larger the numerical value is, the higher the score is, 1-3 commodity images are selected for feature extraction, and about 310000 effective images are selected for each commodity image. In the data preprocessing process, users with the same scores under the same commodity are used as user sequences<u 1 ,u 2 ,u 3 ,…,u x >Here, x is 50, and commodity scores are according to commodity composition sequences of the same scores of the same user<i 1 ,i 2 ,i 3 ,…,i y >Here, taking 20, this procedure resulted in about 200000 pieces of merchandise sequence data, 80% of which were selected, about 160000 pieces for training. The remaining 40000 strips are used as a test set for evaluation and verification of indexes such as accuracy, recall rate and the like.
The user personal information used by the invention mainly comprises user age, user gender, user occupation, region, average score, purchase amount, average browsing time, collection amount and average purchase amount, and totally selects 9 attributes; the commodity attribute information mainly comprises daily order quantity, sales quantity, collection quantity, online time, use season, average service life, average score, price and price interval (three grades are respectively represented by 3, 2 and 1 numbers according to the average price), sales quantity and 9-dimensional attributes. For processing user comments, the invention adopts the ent2Vec tool to extract comment texts into one-dimensional comment vectors with the length of 200, and the ent2Vec tool trains corpus into Chinese wikipedia data. After all comment texts are processed, clustering is carried out on all comment vectors by using a t-means clustering method, and the t-means clustering method is simple and high in efficiency, and scores of end users are totally divided into 30 categories, namely t=70. And processing comments under the commodity by adopting the same method, and clustering after processing to obtain 30 classes. For commodity picture processing, the invention uses VGG16 tool to extract the characteristics of the pictures of the same commodity, if there are multiple pictures, the final characteristic vector is normalized, then t-means clustering is carried out on commodity pictures, and 50 kinds of clustering results are obtained after clustering and are used for distinguishing commodity categories corresponding to the pictures.
The invention uses the grading information of the commodity by the user, the grading information of the user is divided into 5 grades which are 1-5 grades, and the grading vector generator of the user and the grading vector generator of the commodity are constructed based on the grading data of the commodity by the user.
The specific construction process of the user scoring vector generator is as follows: firstly, for the same commodity, extracting the grading information of all users on the commodity, then grouping the grading information according to 1-5 grades, wherein the users with the same grades are in the same group, thus obtaining the grading sequence based on the users as<u 1 ,u 2 ,u 3 ,…u x >Here x takes 50, meaning that 50 users are selected from the group. U for the current user in the sequence r Divide u in predicted sequence r Other user ids maximize the probability product of the other user ids. Since the total number of users w=8000, the loss function y of the scoring user vector generator can be obtained 1 The method comprises the following steps:
Figure BDA0002171116580000161
wherein log () is a logarithmic function with a base of 2, exp () is an exponential function.
After the user scoring vector generator is constructed, the scoring vector of each user can be obtained, and the length is 200.
The specific construction process of the commodity grading vector generator is as follows: firstly, grouping commodities with the same scores of the same user, wherein each group can obtain commodity sequences with the same scores <i 1 ,i 2 ,i 3 ,…,i j >Here j takes 20, representing 20 items from the group. Assume that the current commodity in the sequence is i g The final goal is to predict the division i in the current sequence g The probability of occurrence of other commodities than the commodity total number r=120000, so that the loss function y of the commodity grading vector generator can be obtained by using the maximum likelihood estimation method 2 The method comprises the following steps:
Figure BDA0002171116580000162
/>
after the commodity grading vector generator is constructed, the grading vector of each commodity can be obtained, and the length is 200.
In the second stage of module one (user and item feature extraction), the user feature extraction network takes the user personal information and the user scoring vector obtained in the first stage as input, and jointly trains two tasks: task one is used for predicting user id, task two is used for predicting user comment category.
The 9 attributes of the personal information of the user are respectively represented by one-hot vectors, the obtained 9 one-hot vectors are spliced into a long vector, and the long vector is reduced to a vector with the length of 50 through a five-layer autoencoder tool. Thus, the input to the user feature extraction network is a vector of length equal to 250.
In a user feature extraction network, firstly, convolving a user feature extraction network by adopting 2 convolution layers with the convolution kernel size of 3*1 and the sliding step length of 1; then, a pooling layer is adopted to operate the vector after convolution, and the window size of the pooling layer is 2*1; the convolution 2 layer adopts 4 convolution kernels with the size of 3*1 and the step length of 2 to carry out convolution operation; after batch standardization, the third layer convolution is carried out, the convolution kernel size is 3*1, the step length is 1, and the number of convolution kernels is 5; after batch normalization, the method reaches a fourth layer of convolution, the step length of a fourth layer of convolution kernel is set to be 2, the size of the convolution kernel is 5*1, and the number of the convolution kernels is 10; then the second layer of pooling layer is reached after Dropout, and the window size of the pooling layer is 2*1; then, dropout operation is adopted and then input into layer 5 convolution, the convolution kernel size is 3 x 1, and the step length of 6 convolution kernels is 2; then passing through a third pooling layer, wherein the window size is 4*1; finally, the data are input into a full connection layer with the number of the neurons being 200, and the values of the neurons of 200 form the characteristic vector of the user.
In order to realize task one, the invention accesses the user characteristic vector to a softmax (exponential maximum function) layer after the full connection layer, wherein the softmax layer has 8000 nodes, which represent 8000 users in the system, and the user set formed by the nodes is U. Meanwhile, in order to realize the task II, after the full connection layer, the user feature vector is accessed into another softmax layer, wherein the softmax layer has 30 nodes, 30 user comment categories are represented, and the formed user comment category set is T.
Both tasks employ a softmax multi-class loss function:
1) Let the current user be u r User u can be calculated by using softmax r Is P (u) r ):
Figure BDA0002171116580000171
Wherein sx (u) r ) Is u r Input values for corresponding nodes in the softmax layer.
2) Let the comment category of the current user be z r The user comment category z can be calculated by adopting softmax r Is P (z) r ):
Figure BDA0002171116580000172
Wherein sx (z) r ) Is z r Input values for corresponding nodes in the softmax layer.
The total loss function of the user feature extraction network is thus:
Figure BDA0002171116580000181
wherein eta is L 1 The adjustment coefficient of (2) is between 0.7 and 0.8, here 0.75.
On the other hand, in the second stage of the first module (user and article feature extraction), the article feature extraction network takes the article attribute information and the article scoring vector obtained in the first stage as inputs, and performs joint training on four tasks: task one is used for predicting commodity id, task two is used for predicting commodity category, task three is used for predicting commodity picture category, task four is used for predicting commodity comment category.
The 9 attributes of the commodity attribute information are respectively represented by one-hot vectors, the 9 obtained one-hot vectors are spliced into a long vector, and the long vector is subjected to dimension reduction through a five-layer autoencoder tool to form a vector with the length of 50. Thus, the input to the commodity feature extraction network is a vector of length equal to 250.
In the article characteristic extraction network, the number of convolution kernels of a first layer of convolution layers is 5, the size of the convolution kernels is 3*1, and the step length is 2; then, carrying out second convolution, wherein the size of a convolution kernel of the second layer is 3*1, the number of convolution kernels is 73, and the sliding step length is 3; after batch standardization, the material enters a first pooling layer, and the window size of the pooling layer is 2*1; the method comprises the steps of entering a third layer of convolution after batch standardization, wherein the convolution kernel size is 5*1, the step length is 3, and the number of the convolution kernels is 10; after passing through the Dropout layer, the material enters a second pooling layer, and the window size is 4*1; then, the method enters a 4 th layer convolution layer, the fourth layer convolution parameter is set to be a convolution kernel with the convolution kernel size of 3*1, wherein 5 step sizes are 2; then passing through a pooling layer with a window size of 2*1; then entering a fifth layer of convolution layer through Dropout, wherein the convolution kernel is 3*1 in size, the step length is 3, and the number of the convolution kernels is 6; finally, the data are input into a full-connection layer with the number of the neurons being 200, and the values of the neurons of the 200 form the characteristic vector of the commodity.
In order to realize task one, the invention inserts commodity feature vectors into a first softmax (exponential maximum function) layer after the full connection layer, wherein the softmax layer has 120000 nodes which represent 120000 commodities in the system, and the user set formed by the softmax layer is V. In order to achieve task two, after the fully connected layer, the commodity feature vector is accessed to a second softmax layer, the softmax layer has 15 nodes, 15 commodity categories are represented, and the commodity category set formed is A. In order to realize task three, after the full connection layer, the commodity feature vector is accessed to a third softmax layer, wherein the softmax layer has 50 nodes, which represent 50 commodity picture categories, and the constituted commodity picture category set is I. In order to realize task four, after the full connection layer, the commodity feature vector is accessed to a fourth softmax layer, wherein the softmax layer has 30 nodes, which represent 30 commodity comment categories, and the formed commodity comment category set is M.
All four tasks use softmax multi-class loss functions:
1) Let the current commodity be i x The commodity i can be calculated by adopting softmax x Is P (i) x ):
Figure BDA0002171116580000182
Wherein sx (i) x ) Is i x The input values of the corresponding nodes in the first softmax layer.
2) Let the current commodity category be a x Commodity category a can be calculated by using softmax x Is P (a) x ):
Figure BDA0002171116580000191
Wherein sx (a) x ) Is a as x The input values of the corresponding nodes in the second softmax layer.
3) Let the current commodity picture category be v x The commodity picture category v can be calculated by adopting softmax x Is P (v) x ):
Figure BDA0002171116580000192
Wherein sx (v) x ) V is x The input values of the corresponding nodes in the third softmax layer.
4) Let the current commodity comment category be m x Commodity comment category m can be calculated by adopting softmax x Is P (m) x ):
Figure BDA0002171116580000193
Wherein sx (m) x ) Is m x The input values of the corresponding nodes in the fourth softmax layer.
The overall loss function of the commodity feature extraction network is thus:
Figure BDA0002171116580000194
wherein lambda is 1 ~λ 4 Is L 2 And lambda is the adjustment coefficient of (1) 1234 =1,λ 1 Takes a value of 0.1-0.3, here 0.15 lambda 2 Takes a value of 0.2-0.5, here 0.3 lambda 3 Takes a value of 0.2-0.4, here 0.3 lambda 4 The value is between 0.2 and 0.3, here 0.25.
In the third stage of module one (merchandise sequence recommendation training), the user history merchandise sequence for training includes g=40 merchandise, where there is g input into the encoder 1 =20 products with g-g for decoder output prediction 1 =20 goods. The number of neural networks per GRU unit in the encoder and decoder is set to 1000 and the length of the background vector generated by the encoder is 200.
For the decoder, assume that its output 20 commodity sequences are: < i 1 ,i 2 ,...,i 20 The output of each commodity depends on the output of the previous commodity and also on the background vector c. Let s k (1 is less than or equal to k is less than or equal to 20) is commodity i k Corresponding decoder hidden variable, then s k =GRU(i k-1 ,c,s k-1 ) Thereby commodity i k Output probability p (i) k ) Can be expressed as:
p(i k |i 1 ,i 2 ,...,i k-1 ,c)=p(i k-1 ,s k ,c)。
thus, the joint probability of the maximized output sequence of the kth node of the decoder can be obtained as:
Figure BDA0002171116580000201
thus, the loss function of the decoder output sequence is: y is 3 =-log P(i 1 ,i 2 ,...,i k )。
In order to improve the accuracy of recommendation, the invention adopts a Beam Search strategy in the training process of a decoder, wherein the beam_size is set to be 2, and the specific process is as follows: firstly, when a first commodity is generated, selecting 2 commodities a and b with the highest probability from candidate sequences; then when generating the second commodity, combining the current two commodity sequences < a >, < b > with the 2 commodities with the largest next prediction probability, such as c, d, to obtain four commodity sequences < ac >, < ad >, < bc >, < bd >, and then selecting the sequence with the largest 2 reward (return) values; the above process is repeated until 20 recommended products are finally produced. The Reward value is calculated from the indexes of the three main streams, namely, an NDCG (Normalized Discounted Cumulative Gain: normalized loss cumulative gain) index, a Recall index and a MAP (MeanAverage Precision: average precision average) index.
reward(i 1 ,i 2 ,i k )=0.4×NDCG(i 1 ,i 2 ,i k )+0.4×Recall(i 1 ,i 2 ,i k )+0.2×MAP(i 1 ,i 2 ,i k )。
The total loss function of the encoder can thus be expressed as:
L 3 =-log P(i 1 ,i 2 ,...,i k )-reward(i 1 ,i 2 ,...,i k )。
after the commodity sequence recommendation training phase is completed, the feature vector of each User is stored in a database table user_F, wherein the database table comprises two columns, the first column is a User id, and the second column is a User feature vector. Similarly, the feature vector of each commodity is stored in a database table item_f, which is also in two columns, the first column being the commodity id and the second column being the commodity feature vector.
In the second module (online application of personalized recommendation of commodity sequences), for the current user, the method firstly acquires 20 commodities recently purchased by the user from a purchase log file of a system, forms a commodity sequence, then identifies commodity ids of the 20 commodities, and sequentially acquires feature vectors of the 20 commodities from an item_F database table based on the commodity ids. The feature vectors of the 20 commodities are sequentially input into a trained encoder to generate background vectors, and meanwhile, the feature vectors of the current User are obtained from a user_F database table based on the current User id. On the basis, the background vector and the user characteristic vector are spliced, then the spliced background vector and the user characteristic vector are input into a trained decoder, a commodity sequence formed by 20 commodities is generated, and the sequence is recommended to a current user as an optimal commodity sequence.
The foregoing description is only of the preferred embodiments of the present invention and is not intended to limit the scope of the invention, and all equivalent structural changes made by the description of the present invention and the accompanying drawings or direct/indirect application in other related technical fields are included in the scope of the invention.

Claims (9)

1. The personalized commodity sequence recommending method for the dynamic preference of the user is characterized by comprising the following steps of:
s10, obtaining personal information of a user, commodity attribute information, commodity pictures, scores of the same user corresponding to different commodities, scores of different users of the same commodity, comment texts of the same user on the commodities and comment texts of the same commodity from the scoring information of the user on the commodities;
s20, constructing a user scoring vector generator according to the scoring of different users received by the same commodity, and generating a user scoring vector; constructing a commodity grading vector generator according to the grading corresponding to different commodities by the same user, and generating commodity grading vectors;
s30, inputting comment texts of the same user on all commodities into a user scoring vector generator to generate user comment vectors, and carrying out mean value clustering on the user comment vectors to generate user comment categories; inputting comment texts of users born by the same commodity into a commodity grading vector generator to generate commodity comment vectors, and carrying out mean value clustering on the commodity comment vectors to generate commodity comment categories; extracting picture feature vectors of the commodity by adopting a depth convolution network, and carrying out mean value clustering on the picture feature vectors to generate commodity categories corresponding to the picture;
S40, inputting the personal information of the user and the scoring vector of the user into a user feature extraction network for joint training to obtain a user ID and a user comment prediction type; inputting commodity attribute information and commodity grading vectors into a commodity feature extraction network for joint training to obtain commodity IDs, commodity comment prediction categories and commodity picture prediction categories;
s50, inputting the characteristic vectors of the user characteristic vectors and the characteristic vectors of the historical commodity sequences into a commodity sequence recommendation network, training by a training encoder-decoder in the commodity sequence recommendation network to generate the user characteristic vectors corresponding to each user ID, taking the commodity characteristic vectors corresponding to each commodity ID as personalized recommended commodity sequences, and learning by combining a search strategy to obtain the recommendation of the optimal commodity sequences;
the scoring information of the users used by the user scoring vector generator is divided into 5 groups, and the construction method of the user scoring vector generator comprises the following steps:
grouping into 5 groups according to the user scoring information under the same commodity, wherein the users with the same score are in the same group to obtain a user composition sequence with the same score of the same commodity<u 1 ,u 2 ,u 3 ,…u x >X is 50, which means that 50 users are selected for each group, and the sequence is selected<u 1 ,u 2 ,u 3 ,...,u r ,…u x >As training sample, use user u r As input, divide user u in the output prediction sequence r Other users 'scoring vectors are excluded so that the probability product of other user IDs is maximized, and if the total number of users w=8000, the user scoring vector generator's penalty function y 1 The method comprises the following steps:
Figure FDA0004202432770000021
obtaining the length of a scoring vector of each user to be 200;
wherein u is n,j For the nth user in the j-th group of user sequences, u j A j-th user in the recommendation system;
n is the position number of the current user of the user sequence, j is the position number of the current user in the recommendation system;
the construction method of the commodity grading vector generator comprises the following steps:
dividing the same user with the same scoring commodity into a group, and obtaining the commodity sequence with the same scoring from each group<i 1 ,i 2 ,i 3 ,…,i g ,...,i j >J is 20, which means that 20 commodities are selected in each group, and the commodity sequences are selected<i 1 ,i 2 ,i 3 ,…,i g ,...,i j >As training sample, commodity i g As input, divide i in the output prediction sequence g If the total number of commodities R=120000, obtaining the loss function y of the commodity grading vector generator by using a maximum likelihood estimation method 2 The method comprises the following steps:
Figure FDA0004202432770000022
obtaining the length of a scoring vector of each commodity to be 200;
wherein i is n',j For the nth commodity in the commodity sequence, i j 'is the j' th merchant in the recommendation systemA product;
n 'is the position number of the current commodity of the commodity sequence, and j' is the position number of the current commodity in the recommendation system.
2. The personalized recommendation method for commodity sequences facing user dynamic preferences according to claim 1, wherein said S30 comprises:
the sentence embedded type end 2Vec tool takes Chinese Wikipedia data as a training corpus, extracts each comment text in the training corpus into one-dimensional comment vectors by adopting the sentence embedded type end 2Vec tool, performs t-means clustering on all the comment vectors to obtain N types of user comment categories, extracts one-dimensional commodity comment vectors from the comment text under the same commodity by adopting the sentence embedded type end 2Vec tool, and performs t-means clustering on all the commodity comment vectors to obtain M commodity comment categories; and extracting feature vectors of the commodity pictures by adopting a deep convolution network VGG16, and t-means clustering the comment vectors to obtain I commodity picture categories.
3. The personalized recommendation method for a commodity sequence for a dynamic user preference according to claim 2, wherein said user feature extraction network in S40 comprises:
a first convolution layer adopts 2 convolution kernels with the size of 3*1, and the sliding step length is 1;
the first user pooling layer window size is 2*1;
the second user convolution layer adopts 4 convolution layers with the size of 3*1 and the convolution kernel step length of 2; the third user convolution layer adopts 5 convolution kernels with the size of 3*1, and the sliding step length is 1;
The fourth user convolution layer adopts 10 convolution kernels with the size of 5*1, and the step length of the convolution kernels is 2; the pooling layer window size of the second user pooling layer is 2*1;
the fifth user convolution layer adopts 6 convolution kernels with the size of 3*1, and the step length of the convolution kernels is 2;
the window size of the third user pooling layer is 4*1;
the number of neurons in the user full connectivity layer is 200,
the user randomly deactivates the Dropout layer,
the user personal information and the user scoring vector are convolved through a first user convolution layer, the convolved vector is pooled through a first user pooling layer, then enters a third user convolution layer through batch standardization after being convolved through a second user convolution layer, enters a fourth user convolution layer after being convolved through batch standardization again, enters a user random inactivation Dropout layer after being output by the fourth user convolution layer, enters a second user pooling layer, then enters a user random inactivation Dropout layer for being output, enters a fifth user layer convolution layer, and then enters a user full connection layer after being pooled through the third user pooling layer, and the user characteristic vector is output.
4. The personalized commodity sequence recommending method according to claim 3, wherein the user feature vector is connected to a user index maximum function softmax layer after being output by a full connection layer, the softmax layer has 8000 nodes for representing 8000 users in the system, the user set is U, the user feature vector is connected to a comment index maximum function softmax layer after being output by the full connection layer, the softmax layer has 30 nodes for representing 30 user comment categories, the user comment category set is T, the user index maximum function softmax layer and the comment index maximum function softmax layer both adopt softmax multi-classification loss functions,
Let the current user be u r Obtaining a user through a user index maximum function softmax layer
u r Is P (u) r ):
Figure FDA0004202432770000041
Wherein sx (u) r ) Is u r Input values of corresponding nodes in a user index maximum function softmax layer;
the comment category of the current user is z r Obtaining a user comment category z through a comment index maximum function softmax layer r Is P (z) r ):
Figure FDA0004202432770000042
Wherein sx (z) r ) Is z r Input values of corresponding nodes in the comment index maximum function softmax layer;
the total loss function of the user feature extraction network is:
Figure FDA0004202432770000043
wherein eta is L 1 The value of the regulating coefficient is between 0.7 and 0.8;
u r ' z for each user in the user set U r ' for each comment category in the set of user comment categories T.
5. The personalized recommendation method for commodity sequences facing user dynamic preferences according to claim 1, wherein said user personal information comprises at least the following 9 attributes: the method comprises the steps of expressing each attribute by a single hot vector, then splicing all the attributes into a long vector, and reducing the dimension of the long vector into a vector with the length of 50 through a five-layer self-encoder tool; the commodity attribute information at least comprises the following 9 attributes: daily order volume, sales volume, collection volume, online time, use season, average life, average score, price interval.
6. The personalized recommendation method for commodity sequences facing to user dynamic preferences according to claim 2, wherein said commodity feature extraction network in S40 comprises:
the first commodity convolution layer adopts 5 convolution kernels with the size of 3*1 and the step length of 2;
a second commodity convolution layer which adopts 73 convolution kernels with the size of 3*1, and the sliding step length is 3;
a first commodity pooling layer having a window size of 2*1;
a third commodity convolution layer which adopts 10 convolution kernels with the size of 5*1 and the step length of 3;
a second commodity pooling layer having a window size of 4*1;
a fourth commodity convolution layer which adopts 5 convolution kernels with the step length of 2 and the size of 3*1;
a third commodity pooling layer with a window size of 2*1;
a commodity random inactivation layer;
a fifth commodity convolution layer, which adopts 6 convolution kernels with the size of 3*1 and the step length of 3;
a commodity full-connection layer, the number of the neurons of which is 200,
the commodity attribute information and the commodity grading vector are input into a first convolution layer of a commodity feature extraction network and then are connected into a second convolution layer, the second convolution layer is input into a first pooling layer after output and is input into a third convolution layer after input into a commodity random inactivation layer after input into the second pooling layer after output, the second pooling layer is input into a fourth convolution layer after output, the fourth convolution layer is input into the third pooling layer after output, the third pooling layer is input into the commodity random inactivation layer after output into a fifth convolution layer, the commodity full-connection layer is input after output into the fifth convolution layer, and finally the commodity feature vector is output.
7. The personalized recommendation method for commodity sequences facing to user dynamic preference according to claim 6, wherein the commodity feature vector is output from the commodity full-connection layer and then is connected to a commodity first index maximum function softmax layer, the softmax layer has 120000 nodes, 120000 commodities in the system are represented, the formed user set is V, the commodity feature vector is output from the commodity full-connection layer and then is connected to a commodity second index maximum function softmax layer, the softmax layer has 15 nodes, 15 commodity categories are represented, and the formed commodity category set is A; the commodity feature vector is output from the commodity full-connection layer and then is connected to a commodity third index maximum function softmax layer, the softmax layer is provided with 50 nodes, 50 commodity picture categories are represented, and a formed commodity picture category set is I; the commodity feature vector is output from the commodity full-connection layer and then is connected into a commodity fourth index maximum function softmax layer, the softmax layer is provided with 30 nodes which represent 30 commodity comment categories, the formed commodity comment category set is M,
the commodity first index maximum function softmax layer, the commodity second index maximum function softmax layer, the commodity third index maximum function softmax layer and the commodity fourth index maximum function softmax layer all adopt softmax multi-class loss functions,
Let the current commodity be i x Obtaining commodity i through a commodity first index maximum function softmax layer x Is P (i) x ):
Figure FDA0004202432770000061
Wherein sx (i) x ) Is i x Input values for corresponding nodes in the first softmax layer;
let the current commodity category be a x Obtaining commodity category a through commodity second index maximum function softmax layer x Is P (a) x ):
Figure FDA0004202432770000062
Wherein sx (a) x ) Is a as x Input values for corresponding nodes in the second softmax layer;
let the current commodity picture category be v x Obtaining commodity picture category v through commodity second index maximum function softmax layer x Is P (v) x ):
Figure FDA0004202432770000063
Wherein sx (v) x ) V is x Input values of corresponding nodes in the third softmax layer;
let the current commodity comment category be m x Obtaining commodity comment category m through commodity second index maximum function softmax layer x Is P (m) x ):
Figure FDA0004202432770000071
Wherein sx (m) x ) Is m x The input values of the corresponding nodes in the fourth softmax layer,
the total loss function of the commodity feature extraction network is:
Figure FDA0004202432770000072
wherein lambda is 1 ~λ 4 Is L 2 And lambda is the adjustment coefficient of (1) 1234 =1,λ 1 The value is between 0.1 and 0.3, lambda 2 The value is between 0.2 and 0.5, lambda 3 The value is between 0.2 and 0.4, lambda 4 The value is between 0.2 and 0.3;
v is a set formed by all commodities in the recommendation system, A is a set formed by all commodity categories in the recommendation system, I is a set formed by all commodity picture categories in the recommendation system, and M is a set formed by all commodity comment categories in the recommendation system;
i x ' for each commodity in the commodity set V, a x ' v for each commodity category in commodity category set A x ' for each commodity picture category in commodity picture category set I, m x ' is each commodity comment category in the commodity comment category set M.
8. The personalized recommendation method for commodity sequences facing to user dynamic preference according to claim 1, wherein said S50 specifically comprises:
acquiring g commodities currently purchased by a user, and firstly using the previous g in the g commodities 1 The commodity sequence composed of the individual commodities is used as the input of the encoder and passes g 1 The gating cyclic GRU unit obtains the background vector c of the encoder, then the background vector c is spliced with the user characteristic vector and is used as the input of a decoder, and the decoder is composed of g-g 1 The gate control circulation GRU units are sequentially linked to form output nodes, and the output nodes respectively correspond to the rear g-g in g commodities 1 The number of the neural networks of each gating cycle GRU unit in the encoder and the decoder is set to be 1000, the length of the background vector c generated by the encoder is 200, and c is the background vector of the encoder;
for the decoder, assume that the 20 product sequences output are < i 1 ,i 2 ,...,i 20 >, let s k (1 is less than or equal to k is less than or equal to 20) is commodity i k Corresponding decoder hidden variables s k =GRU(i k-1 ,c,s k-1 ) Then the output probability p (i) k ) The representation is:
p(i k |i 1 ,i 2 ,...,i k-1 ,c)=p(i k-1 ,s k ,c),
the joint probability of the maximized output sequence of the kth node of the decoder is obtained as follows:
Figure FDA0004202432770000081
the loss function of the decoder output sequence is: y is 3 =-logΡ(i 1 ,i 2 ,...,i k ),
Adopting a Beam Search strategy Beam Search, setting the Beam Size beam_Size to be 2, taking an evaluation index as a return, calculating a return value generated by each commodity sequence on each decoding node, selecting the commodity sequence with the largest return value as a current recommended sequence, and calculating the return value through a normalized damage accumulated gain index NDCG, a Recall index Recall and an average precision average value index MAP to obtain the commodity sequence:
reward(i 1 ,i 2 ,i k )=0.4×NDCG(i 1 ,i 2 ,i k )+0.4×Recall(i 1 ,i 2 ,i k )+0.2×MAP(i 1 ,i 2 ,i k ),
the total loss function of the encoder is thus expressed as:
L 3 =-logΡ(i 1 ,i 2 ,...,i k )-reward(i 1 ,i 2 ,...,i k ),
back propagation is carried out according to the error of the recommended sequence, and the weight parameter of the encoder-decoder is adjusted;
the feature vector of each user and the feature vector of each commodity in the commodity sequence recommendation training are stored in a database, the user feature vector of each user is stored in a guide table of the database by taking the user ID as a first column and the corresponding user feature vector as a second column, the commodity feature vector of each commodity is stored in an entry table of the database by taking the commodity ID as a first column and the corresponding commodity feature vector as a second column.
9. The personalized recommendation method for a sequence of items for dynamic preference of a user according to claim 8, further comprising, after S50:
s60, acquiring g currently purchased by the user according to the purchase log file 1 The commodity sequence of each commodity is inquired in a database to obtain the characteristic vector of each commodity in the commodity sequence, and the characteristic vectors are input to an encoder in time sequence to generate a background vector; inquiring a database to obtain a characteristic vector of a current user, splicing a background vector and the characteristic vector of the current user, inputting the spliced background vector and the characteristic vector of the current user into a decoder, and generating g-g 1 And recommending the commodity sequence formed by the commodities to the current user as an optimal commodity sequence.
CN201910763463.6A 2019-08-19 2019-08-19 Commodity sequence personalized recommendation method for dynamic preference of user Active CN110458627B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201910763463.6A CN110458627B (en) 2019-08-19 2019-08-19 Commodity sequence personalized recommendation method for dynamic preference of user

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201910763463.6A CN110458627B (en) 2019-08-19 2019-08-19 Commodity sequence personalized recommendation method for dynamic preference of user

Publications (2)

Publication Number Publication Date
CN110458627A CN110458627A (en) 2019-11-15
CN110458627B true CN110458627B (en) 2023-06-30

Family

ID=68487463

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201910763463.6A Active CN110458627B (en) 2019-08-19 2019-08-19 Commodity sequence personalized recommendation method for dynamic preference of user

Country Status (1)

Country Link
CN (1) CN110458627B (en)

Families Citing this family (14)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110990697A (en) * 2019-11-28 2020-04-10 腾讯科技(深圳)有限公司 Content recommendation method, device, equipment and storage medium
CN111079015B (en) * 2019-12-17 2021-08-31 腾讯科技(深圳)有限公司 Recommendation method and device, computer equipment and storage medium
CN111178946B (en) * 2019-12-17 2023-07-18 清华大学深圳国际研究生院 User behavior characterization method and system
CN111242165B (en) * 2019-12-30 2020-10-23 北京顺达同行科技有限公司 Merchant clustering method and device, computer equipment and storage medium
CN111310038B (en) * 2020-02-06 2022-09-02 腾讯科技(深圳)有限公司 Information recommendation method and device, electronic equipment and computer-readable storage medium
CN111401936B (en) * 2020-02-26 2023-05-26 中国人民解放军战略支援部队信息工程大学 Recommendation method based on comment space and user preference
CN111429234B (en) * 2020-04-16 2023-09-29 电子科技大学中山学院 Commodity sequence recommending method based on deep learning
CN111756807B (en) * 2020-05-28 2021-07-20 珠海格力电器股份有限公司 Multi-split recommendation method and device based on region, storage medium and terminal
CN112070577A (en) * 2020-08-31 2020-12-11 深圳市卡牛科技有限公司 Commodity recommendation method, system, equipment and medium
CN112561644B (en) * 2020-12-16 2023-08-01 深圳市网联安瑞网络科技有限公司 Commodity recommendation method and device based on link prediction and related equipment
CN112612880A (en) * 2020-12-17 2021-04-06 上海自古红蓝人工智能科技有限公司 User-dominated open type intelligent AI corpus submission auditing system and method
CN113781150A (en) * 2021-01-28 2021-12-10 北京沃东天骏信息技术有限公司 Article recommendation method and device
CN112836123B (en) * 2021-02-03 2021-11-16 电子科技大学 Interpretable recommendation system based on knowledge graph
CN114943035A (en) * 2022-06-08 2022-08-26 青岛文达通科技股份有限公司 User dressing recommendation method and system based on self-encoder and memory network

Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109388731A (en) * 2018-08-31 2019-02-26 昆明理工大学 A kind of music recommended method based on deep neural network
CN109408702A (en) * 2018-08-29 2019-03-01 昆明理工大学 A kind of mixed recommendation method based on sparse edge noise reduction autocoding

Family Cites Families (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN106485562B (en) * 2015-09-01 2020-12-04 苏宁云计算有限公司 Commodity information recommendation method and system based on user historical behaviors
CN106294758A (en) * 2016-09-23 2017-01-04 华南师范大学 Collaborative recommendation method based on the change of user cognition degree
CN109241366B (en) * 2018-07-18 2021-10-26 华南师范大学 Hybrid recommendation system and method based on multitask deep learning

Patent Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109408702A (en) * 2018-08-29 2019-03-01 昆明理工大学 A kind of mixed recommendation method based on sparse edge noise reduction autocoding
CN109388731A (en) * 2018-08-31 2019-02-26 昆明理工大学 A kind of music recommended method based on deep neural network

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
"Neural Collaborative Embedding From Reviews for Recommendation";XINGJIE FENG,YUNZE ZENG;《IEEE Access》;103263-103274 *

Also Published As

Publication number Publication date
CN110458627A (en) 2019-11-15

Similar Documents

Publication Publication Date Title
CN110458627B (en) Commodity sequence personalized recommendation method for dynamic preference of user
CN109241424B (en) A kind of recommended method
CN108959603B (en) Personalized recommendation system and method based on deep neural network
CN106920147B (en) Intelligent commodity recommendation method based on word vector data driving
CN111062775A (en) Recommendation system recall method based on attention mechanism
CN104252456B (en) A kind of weight method of estimation, apparatus and system
CN111242729A (en) Serialization recommendation method based on long-term and short-term interests
CN110264291B (en) Information recommendation method and device
CN102193936A (en) Data classification method and device
CN112862569B (en) Product appearance style evaluation method and system based on image and text multi-modal data
CN110737834A (en) Business object recommendation method and device, storage medium and computer equipment
CN112100512A (en) Collaborative filtering recommendation method based on user clustering and project association analysis
CN112182145A (en) Text similarity determination method, device, equipment and storage medium
CN109584006A (en) A kind of cross-platform goods matching method based on depth Matching Model
Malik et al. EPR-ML: E-Commerce Product Recommendation Using NLP and Machine Learning Algorithm
CN111695024A (en) Object evaluation value prediction method and system, and recommendation method and system
CN114861050A (en) Feature fusion recommendation method and system based on neural network
CN112989215B (en) Sparse user behavior data-based knowledge graph enhanced recommendation system
CN116932896A (en) Attention mechanism-based multimode fusion personalized recommendation architecture
CN113722443B (en) Label recommendation method and system integrating text similarity and collaborative filtering
CN114840745A (en) Personalized recommendation method and system based on graph feature learning and deep semantic matching model
CN106919647B (en) Clustering-based network structure similarity recommendation method
CN110020918B (en) Recommendation information generation method and system
CN111339428A (en) Interactive personalized search method based on limited Boltzmann machine drive
Cherednichenko et al. Item Matching Model in E-Commerce: How Users Benefit

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant