WO2023011382A1 - 推荐方法、推荐模型训练方法及相关产品 - Google Patents
推荐方法、推荐模型训练方法及相关产品 Download PDFInfo
- Publication number
- WO2023011382A1 WO2023011382A1 PCT/CN2022/109297 CN2022109297W WO2023011382A1 WO 2023011382 A1 WO2023011382 A1 WO 2023011382A1 CN 2022109297 W CN2022109297 W CN 2022109297W WO 2023011382 A1 WO2023011382 A1 WO 2023011382A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- target
- feature
- data
- samples
- training
- Prior art date
Links
- 238000012549 training Methods 0.000 title claims abstract description 366
- 238000000034 method Methods 0.000 title claims abstract description 102
- 239000013074 reference sample Substances 0.000 claims abstract description 145
- 238000013528 artificial neural network Methods 0.000 claims abstract description 27
- 239000013598 vector Substances 0.000 claims description 407
- 239000000523 sample Substances 0.000 claims description 236
- 238000012545 processing Methods 0.000 claims description 57
- 238000000605 extraction Methods 0.000 claims description 37
- 230000003993 interaction Effects 0.000 claims description 33
- 230000004927 fusion Effects 0.000 claims description 20
- 230000006399 behavior Effects 0.000 description 55
- 230000008569 process Effects 0.000 description 27
- 238000010586 diagram Methods 0.000 description 20
- 238000004891 communication Methods 0.000 description 14
- 238000013473 artificial intelligence Methods 0.000 description 12
- 230000006870 function Effects 0.000 description 10
- 239000011159 matrix material Substances 0.000 description 9
- 238000012360 testing method Methods 0.000 description 9
- 238000004364 calculation method Methods 0.000 description 8
- 238000013500 data storage Methods 0.000 description 7
- 238000013507 mapping Methods 0.000 description 6
- 238000010606 normalization Methods 0.000 description 5
- 230000005540 biological transmission Effects 0.000 description 3
- 230000008878 coupling Effects 0.000 description 3
- 238000010168 coupling process Methods 0.000 description 3
- 238000005859 coupling reaction Methods 0.000 description 3
- 230000006872 improvement Effects 0.000 description 3
- 238000010801 machine learning Methods 0.000 description 3
- MHABMANUFPZXEB-UHFFFAOYSA-N O-demethyl-aloesaponarin I Natural products O=C1C2=CC=CC(O)=C2C(=O)C2=C1C=C(O)C(C(O)=O)=C2C MHABMANUFPZXEB-UHFFFAOYSA-N 0.000 description 2
- 230000009471 action Effects 0.000 description 2
- 230000004913 activation Effects 0.000 description 2
- 238000004422 calculation algorithm Methods 0.000 description 2
- 238000004590 computer program Methods 0.000 description 2
- 238000013135 deep learning Methods 0.000 description 2
- 230000000694 effects Effects 0.000 description 2
- 238000005516 engineering process Methods 0.000 description 2
- 238000002474 experimental method Methods 0.000 description 2
- 238000009434 installation Methods 0.000 description 2
- 230000007246 mechanism Effects 0.000 description 2
- 238000011176 pooling Methods 0.000 description 2
- 230000004044 response Effects 0.000 description 2
- 230000001360 synchronised effect Effects 0.000 description 2
- RPNUMPOLZDHAAY-UHFFFAOYSA-N Diethylenetriamine Chemical compound NCCNCCN RPNUMPOLZDHAAY-UHFFFAOYSA-N 0.000 description 1
- 241000699670 Mus sp. Species 0.000 description 1
- MDIACHVOJQLAOR-ZYDGDJLBSA-N [(7r,8s,9s,10s,13s,14s,17r)-17-ethynyl-10-hydroperoxy-7,13-dimethyl-3-oxo-2,6,7,8,9,11,12,14,15,16-decahydro-1h-cyclopenta[a]phenanthren-17-yl] heptanoate Chemical compound C([C@H]1C)C2=CC(=O)CC[C@]2(OO)[C@@H]2[C@@H]1[C@@H]1CC[C@](C#C)(OC(=O)CCCCCC)[C@@]1(C)CC2 MDIACHVOJQLAOR-ZYDGDJLBSA-N 0.000 description 1
- 230000001133 acceleration Effects 0.000 description 1
- 238000004458 analytical method Methods 0.000 description 1
- 238000013475 authorization Methods 0.000 description 1
- 230000015572 biosynthetic process Effects 0.000 description 1
- 230000000052 comparative effect Effects 0.000 description 1
- 238000011217 control strategy Methods 0.000 description 1
- 238000005314 correlation function Methods 0.000 description 1
- 238000013480 data collection Methods 0.000 description 1
- 238000006073 displacement reaction Methods 0.000 description 1
- 238000011156 evaluation Methods 0.000 description 1
- 230000008570 general process Effects 0.000 description 1
- 238000011478 gradient descent method Methods 0.000 description 1
- 230000002452 interceptive effect Effects 0.000 description 1
- 238000002955 isolation Methods 0.000 description 1
- 238000012886 linear function Methods 0.000 description 1
- 239000007788 liquid Substances 0.000 description 1
- 238000007477 logistic regression Methods 0.000 description 1
- 238000004519 manufacturing process Methods 0.000 description 1
- 230000003287 optical effect Effects 0.000 description 1
- 238000004806 packaging method and process Methods 0.000 description 1
- 230000037361 pathway Effects 0.000 description 1
- 230000008447 perception Effects 0.000 description 1
- 238000007781 pre-processing Methods 0.000 description 1
- 230000002441 reversible effect Effects 0.000 description 1
- 230000033764 rhythmic process Effects 0.000 description 1
- 230000001953 sensory effect Effects 0.000 description 1
- 230000003068 static effect Effects 0.000 description 1
- 238000006467 substitution reaction Methods 0.000 description 1
- 238000012706 support-vector machine Methods 0.000 description 1
- 230000009897 systematic effect Effects 0.000 description 1
- 238000013519 translation Methods 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F16/00—Information retrieval; Database structures therefor; File system structures therefor
- G06F16/90—Details of database functions independent of the retrieved data types
- G06F16/95—Retrieval from the web
- G06F16/953—Querying, e.g. by the use of web search engines
- G06F16/9535—Search customisation based on user profiles and personalisation
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/21—Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
- G06F18/214—Generating training patterns; Bootstrap methods, e.g. bagging or boosting
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
- G06N3/045—Combinations of networks
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06Q—INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES; SYSTEMS OR METHODS SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES, NOT OTHERWISE PROVIDED FOR
- G06Q30/00—Commerce
- G06Q30/02—Marketing; Price estimation or determination; Fundraising
- G06Q30/0201—Market modelling; Market analysis; Collecting market data
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06Q—INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES; SYSTEMS OR METHODS SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES, NOT OTHERWISE PROVIDED FOR
- G06Q30/00—Commerce
- G06Q30/02—Marketing; Price estimation or determination; Fundraising
- G06Q30/0201—Market modelling; Market analysis; Collecting market data
- G06Q30/0202—Market predictions or forecasting for commercial activities
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06Q—INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES; SYSTEMS OR METHODS SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES, NOT OTHERWISE PROVIDED FOR
- G06Q30/00—Commerce
- G06Q30/02—Marketing; Price estimation or determination; Fundraising
- G06Q30/0241—Advertisements
- G06Q30/0251—Targeted advertisements
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06Q—INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES; SYSTEMS OR METHODS SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES, NOT OTHERWISE PROVIDED FOR
- G06Q30/00—Commerce
- G06Q30/02—Marketing; Price estimation or determination; Fundraising
- G06Q30/0241—Advertisements
- G06Q30/0251—Targeted advertisements
- G06Q30/0255—Targeted advertisements based on user history
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06Q—INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES; SYSTEMS OR METHODS SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES, NOT OTHERWISE PROVIDED FOR
- G06Q30/00—Commerce
- G06Q30/02—Marketing; Price estimation or determination; Fundraising
- G06Q30/0241—Advertisements
- G06Q30/0251—Targeted advertisements
- G06Q30/0269—Targeted advertisements based on user profile or attribute
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06Q—INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES; SYSTEMS OR METHODS SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES, NOT OTHERWISE PROVIDED FOR
- G06Q30/00—Commerce
- G06Q30/06—Buying, selling or leasing transactions
- G06Q30/0601—Electronic shopping [e-shopping]
- G06Q30/0631—Item recommendations
Definitions
- the invention relates to the technical field of artificial intelligence, in particular to a recommendation method, a recommendation model training method and related products.
- CTR Click-Through-Rate
- Extracting valuable relationships or patterns from tabular data is critical for machine learning systems to learn accurately.
- how to fully mine the rich information contained between rows and columns of tabular data becomes crucial.
- Early models such as logistic regression, support vector machines, and tree models took row samples as input to make predictions.
- the deep model maps the category features of row samples into embedding vectors, and then uses the feature vector of a single row sample for user behavior prediction.
- feature-interaction-based models and user-sequence-based models have become the mainstream of tabular data modeling.
- the feature interaction-based model is dedicated to interacting the features between the columns of each row sample in the tabular data, so as to fully mine the user sequence features to predict user behavior, and make recommendations based on the predicted user behavior.
- This application provides a recommendation method, a recommendation model training method, and related products. Recommendations are made by fusing feature information of target reference samples to improve recommendation accuracy.
- the embodiment of the present application provides a recommendation method, including: obtaining the data to be predicted; obtaining multiple target reference samples from multiple reference samples according to the similarity between the data to be predicted and multiple reference samples; each reference sample Both the data to be predicted and the data to be predicted include user feature field data and item feature field data.
- the user feature field data of the data to be predicted is used to indicate the target user features
- the item feature field data of the data to be predicted is used to indicate the target item features.
- Each target reference The sample and the data to be predicted have part of the same user feature field data and/or item feature field data; the target feature information of the data to be predicted is obtained according to multiple target reference samples and the data to be predicted; the target feature information includes the first target feature vector group and the second target feature vector group, the first target feature vector group is the data to be predicted after vectorization, and the second target feature vector group is obtained by vectorizing multiple target reference samples; the target feature information is input by The deep neural network DNN obtains the output value; according to the output value, it is determined whether to recommend the target item to the target user.
- the output value may be a probability value, which reflects the probability that the target user operates on the target item.
- target users have different understandings of the probability of operating on the target item.
- the probability that the target user operates on the target item can be understood as the probability that the target user clicks on the application;
- the target item is a song
- the probability that the target user operates on the target item can be It is understood as the probability that the target user likes the song;
- the target item is a commodity
- the probability that the target user operates on the target item can be understood as the probability that the target user purchases the commodity.
- the probability value can be post-processed to obtain an output value. For example, when the probability value is greater than the probability threshold, 1 is used as the output value; when the probability value is less than or equal to the threshold, 0 is used as the output value, where 0 means that the target user will not operate the target item, and 1 means that the target user will operate target item.
- the output value is greater than the threshold, it is determined to recommend the target item to the target user, otherwise it is determined not to recommend the target item to the target user.
- the solution of the present application is applied to a scenario where an item is selected from multiple candidate items for recommendation, the output value corresponding to each candidate item can be obtained; then, the candidate item with the highest output value is recommended to the target user, or, The output values of multiple candidate items are sorted, and the top-ranked (for example, the top ten candidate items) candidate items are recommended to the target user. For example, when recommending songs, the output value of each candidate song in the song library can be obtained, and then the top ten songs with output values are recommended to target users.
- the target feature information obtained in this application also includes the feature information of multiple target reference samples vectorized and fused. Since the target reference sample is selected from multiple reference samples through the similarity between the data to be predicted and multiple reference samples, and has part of the same user feature domain data and/or item feature domain data as the data to be predicted, therefore
- the target reference sample is a reference sample that is similar to the data to be predicted among multiple reference samples, so the user behavior in the target reference sample can provide reference and experience for the prediction of the target user's behavior, so that when using the characteristics of the fusion target reference sample
- the target feature information is used to predict the output value, the predicted output value can be more accurate, and the item recommendation based on this output value improves the accuracy of the recommendation.
- the multiple target reference samples also include label data;
- the second target feature vector group is obtained by vectorizing and merging multiple target reference samples, specifically: the second target feature vector group is a pair-to-many
- the user feature domain data, item feature domain data and label data of a target reference sample are vectorized and fused.
- the user feature data of the target reference sample is used to represent the reference user feature
- the item feature data of the target reference sample is used to represent the reference item feature. Since the target reference sample also carries label data, it refers to the real operation behavior of the reference user on the reference item. Therefore, the second feature vector group contains the real operation behavior of the reference user on the reference item, then when using the target feature information to predict the behavior of the target user, it can be combined with the real operation behavior of the reference user on the reference item to predict the target user’s behavior on the target The operation behavior of the item is used to obtain the output value, so that the accuracy of the predicted output value is higher, thereby improving the accuracy of item recommendation.
- the target feature information also includes a third target feature vector group, the third target feature vector group is obtained by pairwise interaction of the target feature vectors in the first vector group, and the first vector group includes the first A set of target feature vectors and a second set of target feature vectors.
- pairwise interaction is the pairwise interaction of the target feature vectors in the first vector group, but pairwise interaction can be freely performed in practical applications.
- pairwise interactions may be performed on multiple first target feature vectors in the first target feature vector group to obtain multiple third target feature vectors; or, multiple second target feature vectors in the second target feature vector group The target feature vectors are pairwise interacted to obtain multiple third target feature vectors, and so on.
- pairwise interactions are performed on the target feature vectors in the first vector group to obtain a plurality of third target feature vectors, so that the target feature information also includes high-order feature information, that is, the first
- the three-target feature vector can represent the connection between various user behaviors, so using higher-order feature information for behavior prediction can further improve the accuracy of the output value. For example, if a first target feature vector indicates that the user is 28 years old, and another first target feature vector indicates that the user is male, then the third target feature obtained by interacting these two target feature vectors The vector indicates that the user is a 28-year-old male.
- each target feature vector When each target feature vector is used alone for prediction, if the target item meets the needs of 28-year-olds or meets the needs of men, it is considered that the target user has a certain probability to operate the target item, and the output value obtained is generally greater than the probability threshold , and after interacting the target feature vector, only when the target item meets the needs of a 28-year-old male, the target user has a certain probability to operate the target item, and the obtained output value will be greater than the probability threshold. Therefore, the obtained The accuracy of the output value is relatively high, further improving the recommendation accuracy.
- the multiple first target feature vectors in the first target feature vector group are concatenated to obtain the second feature vector of the data to be predicted; the multiple first feature vectors of each target reference sample Splicing is performed to obtain the second eigenvector of each target reference sample, and multiple first eigenvectors of each target reference sample are obtained by vectorizing the target reference sample; obtaining the second eigenvector of each target reference sample and the The similarity between the second eigenvectors of the prediction data; according to the similarity between the second eigenvectors of each target reference sample and the second eigenvectors of the data to be predicted, determine the weight of each target reference sample; according to each The weight of each target reference sample is used to fuse the first feature vectors of multiple target reference samples in the same feature domain to obtain the second target feature vector group.
- the weight of the target reference sample with the highest correlation with the data to be predicted among the multiple target reference samples is maximized, and the second target feature vector obtained through fusion mainly indicates
- the characteristic information is the characteristic information of the target reference sample with the highest degree of correlation.
- the method before obtaining a plurality of target reference samples from the plurality of reference samples according to the similarity between the data to be predicted and the plurality of reference samples, the method further includes: obtaining a plurality of original samples, wherein, Each of the original samples includes user feature domain data and item feature domain data; using multiple user feature domain data and multiple item feature domain data of the data to be predicted as elements, the multiple original The samples are inverted indexed to obtain the multiple reference samples.
- An inverted list for example, the first column in each row in the inverted list is an element, that is, a domain data (user characteristic domain data or item characteristic domain data) under the multiple reference samples, and the second column is a plurality of The reference samples contain the reference samples of the domain data.
- each user feature field data and each item feature field data of the data to be predicted are used as elements, and multiple original samples are indexed to obtain multiple reference samples.
- the reference samples corresponding to each user feature field data and the reference samples corresponding to each item feature field can be indexed; then, the reference samples corresponding to each user feature field data , and the reference samples corresponding to each item feature field are merged and deduplicated to obtain the multiple reference samples.
- the data to be predicted is [U4, LA, Student, L2, cell phone, B3], U4, LA, Student, L2, cell phone, and B3 are all used as query words and obtained from the inverted list shown in Table 2
- the reference sample corresponding to LA is [sample 1, sample 3]
- the reference sample corresponding to Student is [sample 1, sample 2, sample 3]
- the reference sample corresponding to L2 is [sample 3]
- the reference samples are [sample 3, sample 4]
- the reference sample corresponding to B3 is [sample 4].
- merge and deduplicate all reference samples obtained from the inverted list to obtain multiple reference samples, namely [sample 1, sample 2, sample 3, sample 4].
- multiple original samples are sorted by inversion to obtain an inversion list. Due to the use of the inverted list, multiple reference samples can be quickly indexed from multiple original samples by using the inverted list, and some irrelevant original samples can be excluded, so that it is not necessary to calculate the similarity with each original sample , to reduce the computational pressure, quickly screen out the target reference samples, and improve the efficiency of item recommendation.
- an embodiment of the present application provides a method for training a recommendation model.
- the recommendation model includes a feature information extraction network and a deep neural network DNN.
- the method includes: obtaining a plurality of training samples, wherein each training sample includes user feature domain data and Item feature domain data; according to the similarity between the first training sample and a plurality of second training samples, a plurality of target training samples are obtained from a plurality of second training samples, wherein the first training sample is one of the plurality of training samples, A plurality of second training samples are part or all of the plurality of training samples except the first training sample, the user feature domain data of the first training sample is used to indicate the first reference user feature, and the item feature domain data of the first training sample Used to indicate the first reference item feature, the first training sample and each target training sample have part of the same user feature domain data and/or item feature domain data; input the first training sample and multiple target training samples into the feature information Extract the network to obtain the target feature information of the first training sample, wherein the target feature information includes
- the first training sample and multiple target training samples are input into the feature information extraction network of the recommendation model to construct target feature information with richer information, so that the target feature information contains both the feature information of the first training sample, that is, A plurality of fourth target feature vectors also includes feature information fused after vectorization of multiple target training samples, that is, a plurality of fifth target feature vectors, and the target training samples are obtained through the first training sample and the multiple second training samples
- the similarity between is selected from multiple second training samples, so the target training sample is a training sample that is relatively similar to the first training sample, so that when using the target feature information of the first training sample for model training, you can refer to
- the feature information (that is, prior knowledge) fused after vectorization of multiple target training samples is used to predict user behavior and obtain the output value, making the predicted output value more accurate, so that the loss obtained during the training process is relatively small, and the model is easier Convergence; in addition, due to the reference of user characteristic information of multiple target training samples, the model can remember more abundant user characteristic information,
- the fifth target feature vector group is obtained by vectorizing multiple target training samples, specifically: the fifth target feature vector group is obtained from multiple target training samples through a feature information extraction network
- the user feature domain data, item feature domain data and label data are vectorized and fused.
- the target training samples carry label data, because the label data of each target training sample reflects the real operation behavior of the user on the item in each target training sample. Therefore, when using the target feature information to predict the behavior of the target user, it can be combined with the real operation behavior of the user on the item in the target training sample to predict the operation of the first reference user in the first training sample on the first reference item. Probability, so that the accuracy of the predicted output value is higher. Due to the higher accuracy of the predicted output value, the loss obtained during the training process is relatively small, the model training cycle is shortened, and the model convergence speed is improved.
- the target feature information further includes a sixth target feature vector group, and the sixth target feature vector group is obtained by pairwise interaction of the target feature vectors in the second vector group through a feature information extraction network, and the second The set of vectors includes a fourth set of target feature vectors and a fifth set of target feature vectors.
- pairwise interactions are performed on the target feature vectors in the second vector group to obtain a plurality of sixth target feature vectors, so that the target feature information also includes high-order feature information, that is, the first
- the six-target feature vector can represent the high-order features of the first reference user. Therefore, using high-order features for behavior prediction can further improve the prediction accuracy of user behavior and further improve the model convergence speed. For example, when a fourth target feature vector indicates that the user is 28 years old, and another fourth target feature vector indicates that the user is male, then the sixth target feature vector obtained by interacting with these two fourth target feature vectors The target feature vector indicates that the user is a 28-year-old male.
- each fourth target eigenvector When each fourth target eigenvector is used alone for prediction, the item meets the needs of 28-year-olds or meets the needs of men, and the user is considered to have a certain probability to operate the item, and after interacting with the target eigenvectors, only When the item meets the needs of the 28-year-old male, the user has a certain probability to operate the item, thereby improving the prediction accuracy of the user's behavior.
- the fusion includes: splicing multiple fourth target feature vectors in the fourth target feature vector group to obtain the second feature vector of the first training sample;
- the first eigenvectors are spliced to obtain the second eigenvectors of each target training sample, and the multiple first eigenvectors of each target training sample are obtained by vectorizing the target training samples; obtaining the second eigenvectors of each target training sample
- the similarity between the feature vector and the second feature vector of the first training sample according to the similarity between the second feature vector of each target training sample and the second feature vector of the first training sample, determine each target training
- the weight of the sample according to the weight of each target training sample, the first feature vectors of multiple target training samples in the same feature domain are fused to obtain the fifth target feature vector group.
- the weight of the target training sample with the highest degree of correlation with the first training sample among the multiple target training samples is the largest, and the characteristic information mainly indicated by the fifth target feature vector obtained through fusion is the target
- the characteristic information of the training samples so as to use the most relevant target training samples as much as possible to guide the prediction of the behavior of the first reference user, so that the probability accuracy of the predicted first reference user operating on the first reference item is higher , which improves the model convergence speed.
- the method before obtaining a plurality of target training samples from the plurality of second training samples according to the similarity between the first training sample and the plurality of second training samples, the method further includes: The multiple user feature domain data and the multiple item feature domain data of the training samples are used as elements, and the multiple training samples are inverted indexed to obtain the multiple second training samples.
- an inverted list of multiple training samples based on the user feature domain data and item feature domain data of each training sample, where the inverted list contains the correspondence between elements and samples, as shown in Table 2 , the first column in each row in the inverted list is an element, that is, a field data (user feature field data or item feature field data) under the sample, and the second column is the field data contained in multiple reference samples Reference samples.
- each user feature domain data and each item feature domain data in the first training sample as elements to index multiple second training samples from multiple training samples, that is, according to the inverted list Corresponding relationship, the training samples corresponding to each user feature field data and the training samples corresponding to each item feature field can be obtained; then, the training samples corresponding to each user feature field data, and the training samples corresponding to each item feature field The training samples corresponding to domains are combined and deduplicated to obtain the plurality of second training samples.
- multiple training samples are sorted by using the inverted index to obtain an inverted list. Due to the use of the inverted list, multiple second training samples can be quickly found by using the inverted list, without calculating the similarity with each training sample, reducing the computational pressure, and quickly obtaining from multiple second training samples Multiple target training samples to improve model training speed.
- the embodiment of the present application provides a recommendation device, including: an acquisition unit and a processing unit; the acquisition unit is used to acquire the data to be predicted; Obtain multiple target reference samples from multiple reference samples; each reference sample and the data to be predicted include user feature domain data and item feature domain data, and the user feature domain data of the data to be predicted is used to indicate the target user features, and the data to be predicted
- the item feature field data of the data is used to indicate the target item feature, and each target reference sample and the data to be predicted have part of the same user feature field data and/or item feature field data; according to multiple target reference samples and the data to be predicted, the target
- the target feature information of the prediction data includes the first target feature vector group and the second target feature vector group, the first target feature vector group is the data to be predicted after vectorization, and the second target feature vector group is the pair of multiple
- the target reference samples are vectorized and then fused; the target feature information is used as input to obtain the output value through the deep neural network DNN; according to the output value, it
- the multiple target reference samples also include label data;
- the second target feature vector group is obtained by vectorizing and merging multiple target reference samples, specifically: the second target feature vector group is a pair-to-many
- the user feature domain data, item feature domain data and label data of a target reference sample are vectorized and fused.
- the target feature information also includes a third target feature vector group, the third target feature vector group is obtained by pairwise interaction of the target feature vectors in the first vector group, and the first vector group includes the first A set of target feature vectors and a second set of target feature vectors.
- the processing unit is specifically configured to: concatenate a plurality of first target feature vectors in the first target feature vector group to obtain a second feature vector of the data to be predicted ; Concatenate multiple first eigenvectors of each target reference sample to obtain a second eigenvector of each target reference sample, and multiple first eigenvectors of each target reference sample are obtained by vectorizing the target reference sample ; Obtain the similarity between the second eigenvector of each target reference sample and the second eigenvector of the data to be predicted; according to the second eigenvector of each target reference sample and the second eigenvector of the data to be predicted The similarity is to determine the weight of each target reference sample; according to the weight of each target reference sample, the first feature vectors of multiple target reference samples in the same feature domain are fused to obtain the second target feature vector group.
- the processing unit Before the processing unit acquires a plurality of target reference samples from the plurality of reference samples according to the similarity between the data to be predicted and the plurality of reference samples, the processing unit is further configured to: acquire a plurality of original samples, Wherein, each of the original samples includes user feature field data and item feature field data;
- the embodiment of the present application provides a recommendation model training device, the recommendation model includes a feature information extraction network and a deep neural network DNN, the device includes: an acquisition unit and a processing unit; the acquisition unit is used to acquire multiple training samples, Wherein, each training sample includes user feature domain data and item feature domain data; the processing unit is used to obtain multiple target training samples from multiple second training samples according to the similarity between the first training sample and multiple second training samples.
- the first training sample is one of the multiple training samples
- the multiple second training samples are part or all of the multiple training samples except the first training sample
- the user feature domain data of the first training sample is used
- the item feature domain data of the first training sample is used to indicate the first reference item feature
- the first training sample and each target training sample have part of the same user feature domain data and/or item feature domain Data
- the first training sample and a plurality of target training samples are input to the feature information extraction network to obtain the target feature information of the first training sample, wherein the target feature information includes the fourth target feature vector group and the fifth target feature vector group,
- the fourth target feature vector group is obtained by vectorizing the first training sample through the feature information extraction network
- the fifth target feature vector group is obtained by vectorizing multiple target training samples through the feature information extraction network
- the target feature The information is input into the deep neural network DNN to obtain an output value, which is used to represent the probability that the first reference user operates on the first reference item
- the recommendation model is trained according to the output value and the label data
- the fifth target feature vector group is obtained by vectorizing multiple target training samples, specifically: the fifth target feature vector group is obtained from multiple target training samples through a feature information extraction network
- the user feature domain data, item feature domain data and label data are vectorized and fused.
- the target feature information further includes a sixth target feature vector group, and the sixth target feature vector group is obtained by pairwise interaction of the target feature vectors in the second vector group through a feature information extraction network, and the second The set of vectors includes a fourth set of target feature vectors and a fifth set of target feature vectors.
- the processing unit is specifically configured to: concatenate multiple fourth target feature vectors in the fourth target feature vector group to obtain the second feature of the first training sample vector; multiple first feature vectors of each target training sample are spliced to obtain a second feature vector of each target training sample, and multiple first feature vectors of each target training sample are vectorized for the target training sample obtain; obtain the similarity between the second eigenvector of each target training sample and the second eigenvector of the first training sample; according to the second eigenvector of each target training sample and the second eigenvector of the first training sample The similarity between them determines the weight of each target training sample; according to the weight of each target training sample, the first feature vectors of multiple target training samples in the same feature domain are fused to obtain the fifth target feature vector Group.
- the processing unit before the processing unit obtains a plurality of target reference samples from the plurality of reference samples according to the similarity between the data to be predicted and the plurality of reference samples, the processing unit further uses At:
- each user feature domain data and each item feature domain data of each training sample as elements, multiple training samples are inverted indexed to obtain an inverted list; each user feature domain data of the first training sample and The feature field data of each item is used as a query word, and a plurality of second training samples are obtained from the inverted list.
- the processing unit before the processing unit obtains a plurality of target training samples from the plurality of second training samples according to the similarity between the first training sample and the plurality of second training samples, the processing unit , also used in:
- the embodiment of the present application provides an electronic device, including: a memory for storing programs; a processor for executing the programs stored in the memory; when the programs stored in the memory are executed, the processor is used to implement the above-mentioned first The method of the first aspect or the second aspect.
- the embodiment of the present application provides a computer-readable medium, the computer-readable medium stores program code for device execution, and the program code includes the method for realizing the above-mentioned first aspect or the second aspect .
- the embodiment of the present application provides a computer program product containing instructions, and when the computer program product is run on a computer, it enables the computer to implement the method in the first aspect or the second aspect above.
- the embodiment of the present application provides a chip, the chip includes a processor and a data interface, and the processor reads instructions stored in the memory through the data interface to implement the method in the first aspect or the second aspect above.
- the chip may further include a memory, in which instructions are stored, and the processor is used to execute the instructions stored in the memory, and when the instructions are executed, the processor is used to implement the above first aspect or the second aspect method in .
- FIG. 1 is a schematic diagram of an artificial intelligence subject framework provided by an embodiment of the present application.
- FIG. 2 is a schematic diagram of a system architecture provided by an embodiment of the present application.
- FIG. 3 is a chip hardware structure diagram provided by an embodiment of the present application.
- FIG. 4 is a schematic flowchart of a recommendation method provided in the embodiment of the present application.
- FIG. 5 is a schematic diagram of feature vector interaction and splicing provided by the embodiment of the present application.
- Fig. 6 is a structural diagram of a model provided by the embodiment of the present application.
- FIG. 7 is a schematic flowchart of a recommended model training method provided in an embodiment of the present application.
- FIG. 8 is a comparison diagram of a user behavior prediction process provided by the embodiment of the present application.
- FIG. 9 is a schematic diagram of an application recommendation provided by an embodiment of the present application.
- FIG. 10 is a schematic diagram of a product recommendation provided in an embodiment of the present application.
- FIG. 11 is a schematic diagram of a song recommendation provided by an embodiment of the present application.
- FIG. 12 is a structural diagram of a recommendation device provided in an embodiment of the present application.
- FIG. 13 is a structural diagram of a recommended model training device provided in an embodiment of the present application.
- FIG. 14 is a structural diagram of an electronic device provided by an embodiment of the present application.
- FIG. 1 is a schematic diagram of an artificial intelligence subject framework provided by an embodiment of the present application.
- the main framework describes the overall workflow of the artificial intelligence system, which is applicable to the general artificial intelligence field requirements.
- Intelligent information chain reflects a series of processes from data acquisition to processing. For example, it can be the general process of intelligent information perception, intelligent information representation and formation, intelligent reasoning, intelligent decision-making, intelligent execution and output. In this process, the data has undergone a condensed process of "data-information-knowledge-wisdom".
- IT value chain reflects the value brought by artificial intelligence to the information technology industry from the underlying infrastructure of artificial intelligence, information (provided and processed by technology) to the systematic industrial ecological process.
- the infrastructure provides computing power support for the artificial intelligence system, realizes communication with the outside world, and realizes support through the basic platform.
- computing power is provided by smart chips
- smart chips can be central processing unit (central processing unit, CPU), neural network processor (Neural-network Processing Unit, NPU), graphics processing graphics processing unit, abbreviation: GPU), application specific integrated circuit (Application Specific Integrated Circuit, ASIC), field programmable logic gate array (Field Programmable Gate Array, FPGA) and other hardware acceleration chips
- the basic platform includes distributed computing framework and network and other related Platform guarantee and support, which may include cloud storage and computing, interconnection and interoperability network, etc.
- sensors communicate with the outside to obtain data, and these data are provided to the smart chips in the distributed computing system provided by the basic platform for calculation.
- Data from the upper layer of the infrastructure is used to represent data sources in the field of artificial intelligence.
- the data involves graphics, images, voice, text, and IoT data of traditional equipment, including business data of existing systems and sensory data such as force, displacement, liquid level, temperature, and humidity.
- Data processing usually includes data training, machine learning, deep learning, search, reasoning, decision-making, etc.
- machine learning and deep learning can symbolize and formalize intelligent information modeling, extraction, preprocessing, training, etc. of data.
- Reasoning refers to the process of simulating human intelligent reasoning in a computer or intelligent system, using formalized information to carry out machine thinking and solving problems according to reasoning control strategies. Typical functions are search, matching and prediction.
- Decision-making refers to the process of decision-making after intelligent information is reasoned, and usually provides functions such as classification, sorting, and prediction.
- some general-purpose capabilities can be formed based on the results of data processing, such as algorithms or a general-purpose system, such as translation, text analysis, user behavior prediction, computer vision processing, Speech recognition, image recognition, etc.
- Intelligent products and industry applications refer to the products and applications of artificial intelligence systems in various fields. It is the packaging of the overall solution of artificial intelligence, which commercializes intelligent information decision-making and realizes landing applications. Its application fields mainly include: intelligent manufacturing, intelligent transportation, Smart home, smart medical care, smart security, automatic driving, smart terminals, etc.
- FIG. 2 is a schematic diagram of a system architecture 200 provided by an embodiment of the present application.
- the data acquisition device 260 is used to collect multi-domain discrete data including user feature domain data, item feature domain data and label data, that is, training samples, and store the training samples in the database 230.
- the training device 220 is based on the training samples maintained in the database 230 Generate models/rules 201 . The following will describe in more detail how the training device 220 obtains the model/rule 201 based on the training samples.
- the model/rule 201 can process the data to be predicted to obtain the output value, that is, the probability that the target user operates the target item, so as to determine the Whether to recommend the target item to the target user.
- execution device 210 is equipped with I/O interface 212, carries out data interaction with external equipment, "user” can input data to I/O interface 212 through client device 240, for example, can send I/O interface 212 through client device 240
- the /O interface 212 inputs the data to be predicted, wherein the data to be predicted includes user feature field data and item feature field data, and the purpose of the "user" inputting the data to be predicted to the execution device 210 is to obtain the output value, so as to obtain the target user's target The probability that the item performs an action.
- the execution device 210 may call data, codes, etc. stored in the data storage system 250 , and may also store data, instructions, etc. in the data storage system 250 .
- a large number of reference samples are stored in the data storage system 250, and the reference samples can be training samples maintained in the database 230, that is, the database 230 can migrate data to the data storage system 250;
- the correlation function module 213 analyzes the data to be predicted, and inquires out a plurality of target reference samples from the reference samples maintained in the data storage system 250;
- the calculation module 211 uses the model/rule 201 to process the multiple target reference samples and the data to be predicted that are queried by the association function module 213 . Specifically, the calculation module 211 invokes the model/rule 201 to perform vectorization and fusion of multiple target reference samples, and vectorizes the data to be predicted to obtain target feature information of the data to be predicted, and obtains an output value according to the target feature information;
- the calculation module 211 returns the output value to the client device 240 through the I/O interface 212, so that the client device 240 obtains the probability that the target user operates on the target item.
- the training device 220 can generate corresponding models/rules 201 based on different data for different purposes, so as to provide users with better results.
- the user can manually designate the data input into the execution device 210 , for example, operate in the interface provided by the I/O interface 212 .
- the client device 240 can automatically input data to the I/O interface 212 and obtain the result. If the client device 240 needs to obtain authorization from the user for automatically inputting data, the user can set corresponding permissions in the client device 240 .
- the user can view the results output by the execution device 210 on the client device 240, and the specific presentation form may be specific ways such as display, sound, and action.
- the client device 240 may also serve as a data collection terminal and store the collected data into the database 230 .
- FIG. 2 is only a schematic diagram of a system architecture provided by an embodiment of the present invention, and the positional relationship between devices, devices, modules, etc. shown in FIG. 2 does not constitute any limitation.
- the data storage system 250 is an external memory relative to the execution device 210 , and in other cases, the data storage system 250 may also be placed in the execution device 210 .
- FIG. 3 is a hardware structure diagram of a chip provided by an embodiment of the present application.
- a neural network processor (Neural-network Processing Unit, NPU) 30 is mounted on the main central processing unit (Central Processing Unit, CPU) as a coprocessor, and tasks are assigned by the main CPU.
- the core part of the NPU is the operation circuit 303, and the controller 304 controls the operation circuit 303 to extract data in the memory (weight memory 302 or input memory 301) and perform operations.
- the operation circuit 303 includes multiple processing units (Process Engine, PE).
- PE Process Engine
- arithmetic circuit 303 is a two-dimensional systolic array.
- the arithmetic circuit 303 may also be a one-dimensional systolic array or other electronic circuits capable of performing mathematical operations such as multiplication and addition.
- arithmetic circuit 303 is a general-purpose matrix processor.
- the operation circuit 303 fetches the weight matrix B from the weight memory 302 and caches it on each PE in the operation circuit 303 .
- the operation circuit 303 fetches the input matrix A and the weight matrix B from the input memory 301 to perform matrix operations, and the obtained partial or final results of the matrix are stored in an accumulator 308 .
- the vector calculation unit 307 can perform further processing on the output of the operation circuit 303, such as vector multiplication, vector addition, exponential operation, logarithmic operation, size comparison and so on.
- the vector calculation unit 307 can be used for network calculations of non-convolution/non (Fully Connected Layers, FC) layers in neural networks, such as pooling (Pooling), batch normalization (Batch Normalization), and local response normalization (Local Response Normalization) and so on.
- FC Non-convolution/non layers in neural networks, such as pooling (Pooling), batch normalization (Batch Normalization), and local response normalization (Local Response Normalization) and so on.
- vector computation unit 307 stores the processed vectors to unified buffer 306 .
- the vector computing unit 307 may apply a non-linear function to the output of the computing circuit 303, such as a vector of accumulated values, to generate activation values.
- vector computation unit 307 generates normalized values, binned values, or both.
- the processed vectors can be used as activation inputs to arithmetic circuitry 303, eg, for use in subsequent layers in a neural network.
- the operation circuit 303 obtains the data to be predicted from the input memory 301, and obtains the target reference sample from the unified memory 306; then, the operation circuit 303 obtains the data to be predicted according to the data to be predicted and the target reference sample
- the target feature information of the data, and the output value is obtained according to the target feature information, that is, the probability that the target user operates the target item.
- the unified memory 306 is used to store input data (eg, data to be predicted) and output data (eg, output value).
- the storage unit access controller (Direct Memory Access Controller, DMAC) 305 moves the input data in the external memory to the input memory 301 and/or the unified memory 306, stores the weight data in the external memory into the weight memory 302, and stores the weight data in the unified memory
- the data in 306 is stored in the external memory.
- a bus interface unit (Bus Interface Unit, BIU) 310 is used to realize the interaction between the main CPU, DMAC and instruction fetch memory 309 through the bus.
- An instruction fetch buffer 309 for storing instructions used by the controller 304;
- the controller 304 is used for invoking the instructions cached in the instruction fetch memory 309 to control the working process of the operation circuit 303 .
- the unified memory 306, the input memory 301, the weight memory 302 and the instruction fetch memory 309 are all on-chip (On-Chip) memory
- the external memory is a memory outside the NPU
- the external memory can be a double data rate synchronous dynamic random access memory (DDR) Double Data Rate Synchronous Dynamic Random Access Memory, DDR SDRAM), high bandwidth memory (High Bandwidth Memory, HBM) or other readable and writable memory.
- DDR double data rate synchronous dynamic random access memory
- DDR SDRAM Double Data Rate Synchronous Dynamic Random Access Memory
- HBM High Bandwidth Memory
- Tabular Data also known as Multi-Field Categorical Data
- each row in the tabular data is a data point (also called a sample), and each column represents A feature (also called a domain, can also be called a feature domain), each sample contains multiple feature domains; and the value of the sample under each feature domain is called feature domain data, which can also be called domain value.
- a feature also called a domain, can also be called a feature domain
- each sample contains multiple feature domains
- feature domain data which can also be called domain value.
- LA, NYC, LA, and London in Table 1 are the domain values of sample 1, sample 2, sample 3, and sample 4 in the feature domain of the city, respectively.
- the feature domains of each sample include user feature domains and item feature domains, and the domain values under the user feature domain are called user feature domain data, and the domain values under the item feature domain The value is called item feature field data.
- the user feature domain data includes the user's attribute information and the user's behavior sequence (optional), where the user's attribute information includes the user's identification (ID), place of residence, identity, gender, age, etc.
- ID user's identification
- item feature domain data includes item ID, category, trademark, size, color, and other basic information
- user behavior sequence includes user's historical behavior, for example, items that the user has clicked, browsed, and purchased in the past, and so on.
- FIG. 4 is a schematic flowchart of a recommendation method provided by an embodiment of the present application. The method includes the following steps:
- the data to be predicted is multi-domain discrete data, and the data to be predicted includes user feature domain data and item feature domain data.
- the user feature field data of the data to be predicted is used to indicate the target user feature.
- the user feature domain data includes attribute information of the target user, such as the target user's ID, age, gender, place of residence, place of domicile, and other basic information;
- the item feature field data of the data to be predicted is used to indicate the feature of the target item.
- the target item may be a commodity, an application program, a song, a web page, and other items related to the user.
- the feature field data of the target item may have different representations.
- the feature field data of the target item includes the type of application program, installation size, access popularity, installation times, etc.; for another example, if the target item is a song, the feature field data of the target item includes the style of the song , rhythm, duration, playback times, playback popularity, etc.; for another example, the target item is a commodity, and the feature domain data of the target item includes the color, size, price, trademark, manufacturer, evaluation, etc. of the commodity.
- the user feature domain data of the data to be predicted may also include the target user's behavior sequence, for example, the target user's behavior sequence includes items that the target user has clicked, browsed, and purchased in the past, and so on.
- the similarity between the data to be predicted and each reference sample in the multiple reference samples is obtained, and multiple targets are obtained from the multiple reference samples according to the similarity between the data to be predicted and each reference sample Reference samples.
- each reference sample is also multi-domain discrete data, and each reference sample also includes user characteristic domain data and item characteristic domain data.
- the user feature field data of each reference sample is used to indicate the reference user feature
- the item feature field data is used to indicate the reference item feature. Similar to the data to be predicted, the user feature field data of each reference sample contains the attribute information of the reference user, and the item feature field data contains the attribute information of the reference item, such as color, shape, price, and other information, which will not be described here.
- each target reference sample and the data to be predicted have part of the same user feature domain data and/or item feature domain data. It should be noted that in order to ensure that the target reference samples truly serve as a reference for the data to be predicted, it is necessary to make each target reference sample and the data to be predicted have part of the same user feature field data and item feature field data. For example, if the target reference sample and the data to be predicted only have part of the same user feature domain data, for example, both are male, this target reference sample has no reference value in predicting the behavior of the data to be predicted; or, the target reference The sample and the data to be predicted only have part of the same item feature domain data. For example, the purchased items are all black. This kind of target reference sample has no reference value for the behavior prediction of the predicted data. Therefore, in practical applications, compared with the data to be predicted, the obtained target reference samples need to have part of the same user feature domain data and item feature domain data.
- the user feature domain data contained in each reference sample and the user feature domain data contained in the data to be predicted are not exactly the same, while the item feature domain data contained in each reference sample and the item domain feature data contained in the to-be-predicted data can be exactly the same.
- multiple reference samples can be pre-stored in the manner shown in Table 1 to form tabular data, or they can be stored freely, as long as the multiple feature fields of these reference samples are the same as the multiple feature fields of the data to be predicted.
- the plurality of reference samples may be a plurality of original samples in the sample library, or samples selected from a plurality of original samples, wherein each original sample is also multi-domain discrete data, which is different from the to-be-predicted Similar to the data, each original sample also includes user feature domain data and item feature domain data, which will not be described again.
- multiple original samples can be inverted indexed to obtain inverted list to obtain multiple reference samples based on the inverted list.
- each user feature domain data and each item feature domain data of each original sample are used as an element (item), and each original sample is used as a document (document), and multiple original samples are inverted Index to get the inverted list. Since this application only needs to obtain multiple reference samples from multiple original samples, and does not pay attention to the number of occurrences of elements in documents and other information, the posting list in this application may only contain the correspondence between elements and documents.
- each user feature domain data and each item feature domain data of the data to be predicted are taken as elements, and multiple reference samples corresponding to the data to be predicted are indexed from the inverted list. That is, index the reference samples corresponding to each user feature field data of the data to be predicted from the inverted list, and the reference samples corresponding to each item feature field; then merge and deduplicate all the indexed reference samples, Get multiple reference samples. Therefore, compared with the data to be predicted, each reference sample has the same domain data under at least one feature domain, for example, the domain data under the same user feature domain is the same, for example, living in the same city.
- the data to be predicted is [U4, LA, Student, L2, cell phone, B3]
- the corresponding reference sample [sample 1, sample 3], the reference sample corresponding to Student [sample 1, sample 2, sample 3], the reference sample corresponding to L2 is [sample 3], the reference sample corresponding to cell phone [sample 3.
- Sample 4] the reference sample [Sample 4] corresponding to B3. Then, after merging and deduplicating all reference samples, multiple reference samples [sample 1, sample 2, sample 3, sample 4] are obtained.
- the original samples are stored first by inverting, so that a part of the original samples can be indexed from multiple original samples as multiple reference samples, so that only the data to be predicted and multiple reference samples need to be calculated in the future. Similarity, instead of calculating the similarity with multiple original samples, reduces the calculation pressure, so that multiple target reference samples can be quickly obtained.
- the similarity between the data to be predicted and each reference sample can be obtained.
- the similarity between the data to be predicted and each reference sample is obtained through the BM25 algorithm, which will not be described again.
- a reference sample whose similarity is greater than a threshold is used as a target reference sample to obtain multiple target reference samples, or a preset number of reference samples are selected from multiple reference samples in order of similarity from high to low, as Multiple target reference samples.
- the target feature information includes a first target feature vector group and a second target feature vector group, wherein the first target feature vector group is the data to be predicted after vectorization, and the second target feature vector group is the The reference samples are vectorized and then fused.
- the data to be predicted is vectorized to obtain a first target feature vector group, wherein the first target feature vector group includes a plurality of first target feature vectors.
- each user feature field data and each item feature field data of the data to be predicted are encoded to obtain a feature vector of the data to be predicted.
- Coding each user feature field data and each item feature field data of the data to be predicted can be understood as digitizing each user feature field data and each item feature field data of the data to be predicted to obtain The feature vector of the data to be predicted; then, the feature vector of the data to be predicted is mapped to obtain a plurality of target first feature vectors, wherein each first target feature vector is used to represent a feature field data of the data to be predicted, that is, The encoding result of each feature domain data in the data to be predicted is mapped to obtain the first target feature vector corresponding to the feature domain data.
- the data to be predicted contains the behavior sequence of the target user
- the behavior sequence of the target user is encoded, and the encoding result is mapped to obtain the mapping result; then, the mapping result corresponding to the user behavior sequence is fused, A first target feature vector corresponding to the target user's behavior sequence is obtained, and the first target feature vector is used to represent the target user's behavior sequence.
- obtaining multiple first target feature vectors can be achieved through a target recommendation model.
- the training process of the target recommendation model will be described in detail later, and will not be described here.
- the target recommendation model includes a feature information extraction network and a deep neural network (Deep Neural Networks, DNN), wherein, the DNN can be a multi-layer perceptron (Multi-Layer Perceptron, MLP), and in this application, use DNN as MLP An example is used for description, and details are not repeated here.
- DNN Deep Neural Networks
- MLP Multi-Layer Perceptron
- the feature information extraction network includes a coding layer and a mapping layer (embedding layer).
- label data may or may not be carried.
- the following describes the process of obtaining the second target feature vector group with reference to samples with labels and without labels.
- each target reference sample is vectorized to obtain multiple first feature vectors of each target reference sample.
- encode each user feature field data and each item feature field data of each target reference sample to obtain a feature vector of each target reference sample; perform mapping processing on the feature vector of each target reference sample , to obtain multiple first feature vectors of each target reference sample, where each first feature vector is used to represent a feature domain data of the target reference sample.
- the first feature vectors of the plurality of target reference samples are fused to obtain a second target feature vector group, wherein the second target feature vector group includes a plurality of second target feature vectors.
- the multiple first target feature vectors of the data to be predicted and the multiple first feature vectors of each target reference sample determine the weight corresponding to each target reference sample; according to the multiple weights of multiple target reference samples , fusing the first feature vectors of multiple target reference samples in the same feature domain to obtain multiple second target feature vectors.
- multiple first target feature vectors of the data to be predicted are spliced to obtain a second feature vector of the data to be predicted; multiple first feature vectors of each target reference sample are spliced to obtain each target reference sample the second eigenvector of the data to be predicted; obtain the similarity between the second eigenvector of the data to be predicted and the second eigenvector of each target reference sample, and obtain multiple similarities corresponding to multiple target reference samples, wherein the similarity It can be Euclidean distance, cosine similarity, etc.; then, normalize the multiple similarities of multiple target reference samples, and use the normalization result corresponding to each target reference sample as each target reference The weight of the sample. Therefore, the weight of the i-th target reference sample in the target reference sample can be expressed by formula (1):
- a i is the weight of the i-th target reference sample
- q is the second feature vector of the data to be predicted
- r i is the second feature vector of the i-th target reference sample
- similarity(q, e i ) is the i-th
- k is the number of multiple target reference samples.
- the first feature vectors of multiple target reference samples under any same feature domain (that is, each target reference sample corresponds to a first feature vector under the feature domain)
- the jth second target feature vector among the plurality of second target feature vectors can be represented by formula (2):
- R j is the j-th second target feature vector
- e ij is the j-th first feature vector of the i-th target reference sample
- the value of j is an integer from 1 to n
- n is each target reference sample
- the number of multiple first feature vectors of that is, the number of feature fields of each target reference sample, is also the number of multiple second target feature vectors.
- a second feature vector of each target reference sample is obtained. Therefore, after obtaining the second eigenvector of each target reference sample, the weight of each target reference sample can be used to directly fuse multiple second eigenvectors of multiple target reference samples, that is, weighting processing, to obtain a fused Then, according to the reverse order of splicing the multiple first feature vectors of each target reference sample, the fused feature vectors are split to obtain multiple second target feature vectors.
- the order of splicing the multiple first feature vectors is not limited, but it is necessary to ensure that the data to be predicted
- the splicing order of multiple first target feature vectors of is consistent with the splicing sequence of multiple first feature vectors of each target reference sample.
- acquiring multiple second target feature vectors may also be implemented through the above-mentioned target recommendation model.
- the user feature domain data and item feature domain data of each target reference sample are input to the encoding layer to encode each target reference sample to obtain the feature vector of each target reference sample, for example, the i-th target reference
- the feature vector of the sample is (r i1 ,r i2 ,r i3 ,...,r in ), the value of i is from 1 to k, and k is the number of multiple target reference samples; then, the feature of each target reference sample
- the vector is input to the embedding layer to map the eigenvectors of each target reference vector to obtain multiple first eigenvectors of each target reference sample.
- the multiple first eigenvectors of the i-th target reference sample are (e i1 ,e i2 ,e i3 ,...,e in );
- the feature information extraction network also includes an attention layer, which combines multiple first feature vectors of each target reference sample and multiple first feature vectors of the data to be predicted
- the target feature vector (e 1 , e 2 , e 3 ,...,e n ) is input to the attention layer, and the (e 1 ,e 2 ,e 3 ,...,e n ) are spliced to obtain the second feature of the data to be predicted vector, ie Splicing the first feature vectors of each target reference sample to obtain the second feature vector of each target reference sample, for example, the second feature vector of the i-th target reference sample is Then, based on the second eigenvector of the data to be predicted and the second eigenvector of each target reference sample, determine the weight of each target reference sample;
- a eigenvector is
- the multiple second target feature vectors can also be simplified as: (e n+ 1 ,e n+2 ,e n+3 ,...,e 2n ).
- each reference sample also carries tag data, which is used to characterize the actual operation of the reference item by the reference user in the reference sample, for example, when the reference item is an application program, the tag is used to characterize the reference user Whether the application was clicked. Therefore, in the process of vectorizing each target reference sample to obtain multiple first feature vectors of each target reference sample, in addition to converting each user feature domain data and each item feature domain data of each target reference sample In addition to vectorization, the label data of each target reference sample is also vectorized synchronously to obtain multiple first eigenvectors of each target reference sample.
- the plurality of first target feature vectors of each target reference sample obtained by vectorization further includes a first target feature vector used to indicate label data.
- each user feature field data, each item feature field data, and tag data of each target reference sample are encoded to obtain a feature vector of each target reference sample.
- the feature vector of the i-th target reference sample is (r i1 ,R i2 ,r i3 ,...,r in ,r i(n+1) ), where r i(n+1) is the i-th target reference
- An encoding result of the label data of the sample then, map the feature vectors of each target reference sample to obtain a plurality of first feature vectors of each target reference sample.
- the multiple feature vectors of the i-th target reference sample are (e i1 ,e i2 ,e i3 ,...,e in ,e i(n+1) ), where e i(n+1) is used to indicate Label data of the i-th target reference sample.
- the first feature of multiple target reference samples under the same feature domain including user feature domain, item feature domain and tag domain
- Vectors are fused to obtain the second target feature vector group, that is, compared with the above-mentioned case of not carrying label data
- the multiple second feature vectors in the second feature vector group obtained by fusion at this time also contain information for indicating the fused
- the second target feature vector of the labeled data is (e n+1 , e n+2 , e n+3 ,..., e 2n , e 2n+1 ), where e 2n+1 is used to indicate multiple targets
- the fused label data of the reference sample is (e n+1 , e n+2 , e n+3 ,..., e 2n , e 2n+1 ), where e 2n+1 is used to indicate multiple targets
- the fused label data of the reference sample is (e n+1 , e n+2 , e n+3 ,..., e
- the first target feature vector group and the second target feature vector group can be Splicing is performed to obtain the target feature information, then the target feature information is (e 1 , e 2 , e 3 ..., e n , e n+1 , e n+2 , e n+3 ,..., e 2n ) or (e 1 ,e 2 ,e 3 ...,e n ,e n+1 ,e n+2 ,e n+3 ,...,e 2n ,e 2n+1 ); group and the second target feature vector group are concatenated.
- both the first target feature vector group and the second target feature vector group may be used as input data to perform subsequent output value prediction to obtain output values.
- the target feature after obtaining the first target feature vector group and the second target feature vector group, in addition to splicing the first target feature vector group and the second target feature vector group, the target feature can also be Vectors are interacted to obtain high-level feature information.
- multiple first target feature vectors and multiple second target feature vectors can be concatenated (concat) to obtain the first vector group; then, the target in the first vector group The feature vectors are interacted in pairs to obtain the third target feature vector group; then, the first vector group and the third target feature vector group are spliced to obtain target feature information.
- a plurality of third target feature vectors can be represented by formula (3):
- the value of i is 1 to 2n
- the value of j is 2 to 2n
- the value of j is greater than i
- 2n is the number of target feature vectors in the first vector group
- inter is the interactive operation between vectors .
- the pairwise interaction of the above-mentioned vectors is mainly to fuse two vectors into one vector, and the feature information represented by one vector after fusion is the feature information after fusion of the feature information represented by the two vectors.
- the two-to-two interaction of vectors can be realized by vector dot product, kernel product and network layer. The present application does not limit the interaction mode of the two vectors, as long as one vector obtained after fusion can represent the feature information represented by the two vectors.
- the above only shows the case of pairwise interaction of eigenvectors, and in practical applications, three eigenvectors or a larger number of eigenvectors may also be interacted with.
- all target feature vectors in the first vector group are pairwise interacted.
- some target feature vectors can also be selected from the first vector group for interaction. For example, only some of the multiple first target feature vectors and some of the multiple target second feature vectors in the target feature information may be interacted to obtain multiple third target feature vectors. Therefore, the present application does not limit the source of the interacting vectors and the quantity of the interacting vectors.
- the target feature information is input into the deep neural network DNN as input data to obtain an output value.
- the output value is a probability value, which represents the probability that the target user operates on the target item.
- target users have different understandings of the probability of operating on the target item.
- the target item is an application
- the probability that the target user operates on the target item can be understood as the target user’s click probability on the application
- the target item is a song
- the probability that the target user operates on the target item It can be understood as the probability that the target user likes the song
- the target item is a commodity
- the probability that the target user operates on the target item can be understood as the probability that the target user purchases the commodity.
- the probability value can be post-processed to obtain an output value. For example, when the probability value is greater than the probability threshold, 1 is used as the output value, and when the probability value is less than or equal to the threshold value, 0 is used. As an output value, 0 indicates that the target user will not operate the target item, and 1 indicates that the target user will operate the target item.
- the output value when the output value is represented by binary data of 0 or 1, then when the output value is 1, it is determined to recommend the target item to the target user; when the output value is 0, it is determined not to recommend the target item to the target user thing.
- the output value when the output value is expressed in the form of probability, when the probability is greater than the probability threshold, it is determined to recommend the target item to the target user, and when the probability is less than or equal to the probability threshold, it is determined not to recommend the target item to the target user.
- the recommendation method of this application when the recommendation method of this application is applied to a multi-item recommendation scenario, it is necessary to calculate the target user's operation probability for each candidate item; then, sort the operation probabilities of multiple candidate items, and rank the top Candidate items are recommended to target users. For example, when recommending songs, it is necessary to calculate the target user's liking probability for each candidate song, and then recommend the song with the highest liking probability to the target user.
- the acquired target feature information in addition to the feature information of the data to be predicted, also includes feature information obtained by fusion of multiple target reference samples after vectorization. Since the target reference sample and the data to be predicted have part of the same user feature domain data and/or item feature domain data, the user behavior in the target reference sample can provide reference and experience for the prediction of the target user's behavior, so that when using such When the target feature information is used to predict the output value, the predicted output value can be more accurate, and the item recommendation based on this output value improves the accuracy of the recommendation.
- the following describes the process of obtaining the output value in combination with the specific model structure and the way that the reference sample carries label data and interacts with the target feature vector.
- the model includes a feature information extraction network and an MLP, where the feature information extraction network includes a coding layer, an embedding layer, an attention layer and an interaction layer.
- the interaction layer is optional.
- the interaction layer needs to be designed; if the target feature vector is not interacted, the interaction layer need not be designed.
- the target feature information is input to the multi-layer perceptron MLP to obtain the output value.
- FIG. 7 is a schematic flowchart of a method for training a recommendation model provided by an embodiment of the present application.
- the recommendation model includes feature information extraction network and multi-layer perceptron MLP. The method includes the following steps:
- each training sample is multi-domain discrete data, similar to the reference sample above, each training sample includes user feature domain data and user feature domain data. It should be understood that each training sample also carries label data, and the label data of each training sample is used to represent the actual operation of the items in the training sample by the user in each training sample. For example, when the item is an application program, the actual operation status is whether the user clicks on the application program.
- multiple reference samples may be the multiple training samples, or some of the multiple training samples, for example, select some training samples with high data integrity from multiple training samples as a reference sample.
- 702 Acquire multiple target training samples from multiple second training samples according to the similarity between the first training sample and multiple second training samples.
- the first training sample is any one of multiple training samples
- the user feature domain data of the first training sample is used to indicate the first reference user feature
- the item feature domain data of the first training sample is used to indicate the first reference item feature.
- the multiple second training samples are part or all of the multiple training samples except the first training samples.
- the first training sample and each target training sample have part of the same user feature domain data and/or item feature domain data.
- each target reference sample and the first training sample have part of the same user feature domain data and item feature domain data.
- multiple target training samples are acquired from multiple second training samples according to the similarity between the first training sample and each second training sample.
- the second training sample whose similarity is greater than the threshold can be used as the target training sample to obtain multiple target training samples, or a preset number of second training samples can be selected from multiple second training samples in order of similarity from high to low. Two training samples, as multiple target training samples.
- all the training samples except the first training samples in the multiple training samples can be directly used as multiple second training samples, and then the similarity between the first training samples and the multiple second training samples can be obtained; Parts may be selected from training samples other than the first training samples as multiple second training samples according to the manner of the above-mentioned inverted index.
- each user feature domain data and each item feature domain data of each training sample are used as elements, each training sample is used as a document, and multiple training samples are inverted Arrange the index to obtain the inverted list; then, use each user feature field data and each item feature field data of the first training sample as query words, and obtain a plurality of second training samples from the inverted index, therefore, Compared with the first training samples, these second training samples have at least the same domain data under the same feature domain. Therefore, when a certain training sample has different domain data under any same feature domain than the first training sample, this training sample is not used as the second training sample. Therefore, the above-mentioned second training sample may be a part of the multiple training samples other than the first training sample.
- 703 Input the first training sample and multiple target training samples into the feature information extraction network to obtain target feature information of the first training sample.
- the first training sample and each target training sample are input into the feature information extraction network as input data to obtain the target feature information of the first training sample.
- the target feature information includes a fourth target feature vector group (including multiple fourth target feature vectors) and a fifth target feature vector group (including multiple fifth target feature vectors), wherein the multiple fourth target
- the eigenvectors are similar to the above method of obtaining multiple first target eigenvectors, that is, the eigenvectors of the first training samples are mapped to obtain multiple fourth target eigenvectors, wherein the eigenvectors of the first training samples are corresponding to the first
- Each user feature field data and each item feature field data of the training sample are obtained by encoding, and will not be described in detail; optionally, obtaining multiple fifth target feature vectors is the same as the above-mentioned way of obtaining multiple second target feature vectors Similarly, it is obtained by fusing multiple first feature vectors of multiple target training samples in the same feature domain, where the multiple first feature vectors corresponding to each target training sample are the features of each target training sample Vectors are mapped, and the feature vector of each target training sample is obtained by en
- the label data of the target reference sample may not be vectorized and fused, or the label data of the target reference sample may be vectorized and fused .
- the target feature information may also include a sixth target feature vector group (including a plurality of sixth target feature vectors), wherein the acquisition method of the multiple sixth target feature vectors is the same as the acquisition of the above-mentioned multiple third target feature vectors
- the method is similar, that is, splicing multiple fourth target feature vectors and multiple fifth target feature vectors to obtain a second vector group; then, pairwise interaction is performed on the target feature vectors in the second vector group to obtain multiple sixth
- the target feature vector is not described in detail.
- the target feature information of the first training sample is input to the multi-layer perceptron of the recommendation model to obtain an output value, that is, to predict the operation of the first reference user on the first reference item.
- the loss is determined according to the output value and the label data of the first training sample, that is, according to the predicted situation of the first reference user operating the first reference item and the actual situation of the first reference user operating the first reference item. According to the situation, determine the loss; adjust the model parameters of the recommendation model to be trained according to the loss and the gradient descent method, train the recommendation training model, and obtain the target recommendation model.
- the training of the recommendation model is to use multiple training samples for iterative training, wherein the training process of each training sample is similar to the training process using the first training sample shown in Figure 7, and will not be described again; until the recommendation model When converging, the training of the recommendation model is completed and the target recommendation model is obtained.
- the existing item recommendation process is to first use the training samples for model training (consistent with the existing supervised training method), after the model training is completed, the data to be predicted is directly input into the recommendation model for user Behavior prediction, obtain the output value, determine whether to recommend items to the user based on the output value; and the user behavior prediction of the present application is to first use the training samples to train the recommended model (consistent with the training method shown in Figure 7), and the model to be completed After training, use the training sample as a reference sample; when the data to be predicted is obtained, first obtain the target training sample corresponding to the data to be predicted from the training sample, and then input the data to be predicted and the target training sample into the target recommendation model The user behavior prediction is carried out, and the output value is obtained, and based on the output value, it is determined whether to recommend items to the user
- the recommendation of the application for example, the application recommendation under the high-quality application, the application recommendation under the high-quality new tour list; for each type of application recommendation, first obtain multiple candidate applications; Then, based on the user feature field data of the target user and the item feature field data of each candidate application program, the data to be predicted corresponding to each candidate application program is constructed, that is, the user feature field data of the target user and the item feature field data of each candidate application program The data is spliced into a data to be predicted; then, based on the above-mentioned recommendation method and the data to be predicted, the target user's click probability for each candidate application is predicted; then, according to the order of the click probability of each candidate application from high to low, multiple Candidate applications are sorted, and multiple candidate applications are displayed on the recommendation page in order of click probability from high to low or only the top candidate applications are displayed.
- the feature information of the target reference sample will be fused, so that the predicted click rate will be more accurate, and the application program recommended to the target user will be more accurate, thereby increasing the download rate of the application program.
- the data to be predicted corresponding to each candidate commodity can be constructed based on the item feature domain data of each candidate commodity and the user feature domain data of the target user, that is, the target user’s user feature domain data and
- the item feature domain data of each candidate commodity is spliced into a data to be predicted; based on the above-mentioned recommendation method and the data to be predicted, the target user’s purchase probability for each candidate commodity is predicted; then, according to the purchase probability of each candidate commodity from high to high Sort multiple candidate products in descending order, and display the sorted multiple candidate products or only display the top-ranked candidate products in order of purchase probability from high to low on the recommendation page.
- the predicted purchase probability is more accurate, and the products recommended to the target users are more in line with the user's needs, and the sales of the products are increased.
- song recommendation for example, song recommendation in private FM, song recommendation in 30 songs per day; for each type of recommendation, first obtain a plurality of candidate songs, according to the item characteristics of each candidate song domain data and the user feature domain data of the target user, constructing data to be predicted corresponding to each candidate song, that is, splicing the user feature domain data of the target user and the item feature domain data of each candidate song into one data to be predicted; based on the above
- the recommendation method and the data to be predicted predict the score of each candidate song, and the score of each candidate song is used to represent the target user's preference for the candidate song; then, according to the order of the score of each candidate song from high to low Sort multiple candidate songs, and display the sorted multiple candidate songs in descending order of ratings or only display candidate songs with higher ratings on the recommendation page.
- the feature information of the target reference sample will be fused, so that the predicted score is more accurate, and the recommended song is more in line with the user's needs, improving the accuracy of song recommendation.
- Experimental setting 1 Use the following test indicators to evaluate the pros and cons of the prediction accuracy of the model, namely:
- test index AUC the area under the receiver operating characteristic curve and the coordinate axis (Area Under Curve, AUC), loss (Logloss, LL) and relative improvement (relative improvement, REI.Impr); among them, for the test index AUC, the value The larger the effect of the model, the better; for the test index LL, the smaller the value, the better the effect of the model; wherein, the test REI.Impr is the improvement of the prediction accuracy of the model (RIM) of the present application relative to other models, For the test index REI.Impr, the larger the value, the higher the prediction accuracy of RIM relative to the accuracy of the model being compared.
- RIM prediction accuracy of the model
- Experimental setting 2 Obtain the data set of application A, the data set of application B and the data set of application C, and test the CTR based on user behavior prediction on the data set of application A, the data set of application B and the data set of application C respectively.
- Experimental setting 3 Obtain the first data set and the second data set, test the AUC and LL of the model based on feature interaction prediction CTR on the first data set and the second data set, and test the model of this application when predicting CTR AUC and LL.
- the first data set may be avazu and the second data set may be criteo.
- the models for predicting CTR based on user behavior include: HPMN, MIMN, DIN, DIEN, SIM, UBR; the models for predicting CTR based on feature interaction include: LR, GBDT, FM, FFM, AFM, FNN, DeepFM, IPNN, PIN, xDeepFM, FGCNN.
- Table 3 and Table 4 are comparative results.
- FIG. 12 is a structural diagram of a recommendation device provided in an embodiment of the present application.
- the recommendation device 1200 includes an acquisition unit 1201 and a processing unit 1202;
- the processing unit 1202 is configured to obtain the data to be predicted; obtain multiple target reference samples from multiple reference samples according to the similarity between the data to be predicted and the multiple reference samples; each reference sample and the data to be predicted include user feature domain data and item feature field data, the user feature field data of the data to be predicted is used to indicate the target user features, the item feature field data of the data to be predicted is used to indicate the target item features, each target reference sample and the data to be predicted have part of the same user Feature domain data and/or item feature domain data; obtain target feature information of the data to be predicted according to multiple target reference samples and data to be predicted; target feature information includes a first target feature vector group and a second target feature vector group, the first The target feature vector group is the data to be predicted after vectorization, and the second target feature vector group is obtained by vectorizing and merging multiple target reference samples; the target feature information is used as input to obtain the output value through the deep neural network DNN; according to the output The value determines whether to recommend the target item to the target user.
- FIG. 13 is a structural diagram of a recommendation model training device provided by an embodiment of the present application.
- the recommendation model includes feature information extraction network and deep neural network DNN.
- the recommended model training device 1300 includes an acquisition unit 1301 and a processing unit 1302;
- An acquisition unit 1301, configured to acquire a plurality of training samples, wherein each training sample includes user feature domain data and item feature domain data;
- a processing unit 1302 configured to acquire a plurality of target training samples from the plurality of second training samples according to the similarity between the first training sample and the plurality of second training samples, wherein the first training sample is one of the plurality of training samples , a plurality of second training samples is a part or all of a plurality of training samples except the first training sample, the user feature domain data of the first training sample is used to indicate the first reference user feature, and the item feature domain of the first training sample The data is used to indicate the characteristics of the first reference item, and the first training sample and each target training sample have part of the same user feature domain data and/or item feature domain data;
- the target feature information includes the fourth target feature vector group and the fifth target feature vector group
- the fourth The target feature vector group is obtained by vectorizing the first training sample through the feature information extraction network
- the fifth target feature vector group is obtained after vectorizing multiple target training samples through the feature information extraction network
- the target feature information is input into the deep neural network DNN to obtain an output value, and the output value is used to represent the probability that the first reference user operates on the first reference item;
- the recommendation model is trained according to the output value and the label of the first training sample to obtain a target recommendation model.
- FIG. 14 is a structural diagram of an electronic device provided by an embodiment of the present application.
- the electronic device 1400 may be the above-mentioned recommendation device 1200; or, a chip or a chip system in the recommendation device 1200; the electronic device may also be the above-mentioned recommendation model training device 1300; or, a chip or a chip in the recommendation model training device 1300 system;
- the electronic device 1400 includes a memory 1401 , a processor 1402 , a communication interface 1403 and a bus 1404 .
- the memory 1401 , the processor 1402 , and the communication interface 1403 are connected to each other through a bus 1404 .
- the memory 1401 may be a read-only memory (Read Only Memory, ROM), a static storage device, a dynamic storage device or a random access memory (Random Access Memory, RAM).
- the memory 1401 may store programs, and when the programs stored in the memory 1401 are executed by the processor 1402, the processor 1402 and the communication interface 1403 are used to execute various steps in the data stream transmission method of the embodiment of the present application.
- the processor 1402 may be a general-purpose central processing unit (Central Processing Unit, CPU), a microprocessor, an application-specific integrated circuit (Application Specific Integrated Circuit, ASIC), a graphics processing unit (graphics processing unit, GPU) or one or more
- the integrated circuit is used to execute related programs to realize the functions required by the units in the audio feature compensation device or the audio recognition device of the embodiment of the present application, or to execute the data stream transmission method of the method embodiment of the present application.
- the processor 1402 may also be an integrated circuit chip with signal processing capabilities. During implementation, each step in the data stream transmission method of the present application may be completed by an integrated logic circuit of hardware in the processor 1402 or instructions in the form of software.
- the above-mentioned processor 1402 can also be a general-purpose processor, a digital signal processor (Digital Signal Processing, DSP), an application-specific integrated circuit (ASIC), a ready-made programmable gate array (Field Programmable Gate Array, FPGA) or other programmable logic devices , discrete gate or transistor logic devices, discrete hardware components.
- DSP Digital Signal Processing
- ASIC application-specific integrated circuit
- FPGA Field Programmable Gate Array
- Various methods, steps, and logic block diagrams disclosed in the embodiments of the present application may be implemented or executed.
- a general-purpose processor may be a microprocessor, or the processor may be any conventional processor, or the like.
- the steps of the method disclosed in connection with the embodiments of the present application may be directly implemented by a hardware decoding processor, or implemented by a combination of hardware and software modules in the decoding processor.
- the software module can be located in a mature storage medium in the field such as random access memory, flash memory, read-only memory, programmable read-only memory or electrically erasable programmable memory, register.
- the storage medium is located in the memory 1401, and the processor 1402 reads the information in the memory 1401, and combines its hardware to complete the functions required by the units included in the user equipment or the head-mounted device of the embodiment of the present application, or execute the method embodiment of the present application The various steps in the data streaming method.
- the communication interface 1403 can be a transceiver device such as a transceiver to realize communication between the electronic device 1400 and other devices or communication networks; the communication interface 1403 can also be an input-output interface to realize communication between the electronic device 1400 and the input-output device.
- the input-output devices include but are not limited to keyboards, mice, display screens, U disks and hard disks.
- the processor 1402 may acquire the data to be predicted through the communication interface 1403 .
- the bus 1404 may include pathways for transferring information between various components of the device electronics 1400 (eg, memory 1401 , processor 1402 , communication interface 1403 ).
- the electronic device 1400 shown in FIG. 14 only shows a memory, a processor, and a communication interface, in a specific implementation process, those skilled in the art should understand that the electronic device 1400 also includes other necessary components for normal operation. device. Meanwhile, according to specific needs, those skilled in the art should understand that the electronic device 1400 may also include hardware devices for implementing other additional functions. In addition, those skilled in the art should understand that the electronic device 1400 may only include components necessary to realize the embodiment of the present application, and does not necessarily include all the components shown in FIG. 14 .
- the disclosed systems, devices and methods may be implemented in other ways.
- the device embodiments described above are only illustrative.
- the division of the units is only a logical function division. In actual implementation, there may be other division methods.
- multiple units or components can be combined or May be integrated into another system, or some features may be ignored, or not implemented.
- the mutual coupling or direct coupling or communication connection shown or discussed may be through some interfaces, and the indirect coupling or communication connection of devices or units may be in electrical, mechanical or other forms.
- the units described as separate components may or may not be physically separated, and the components shown as units may or may not be physical units, that is, they may be located in one place, or may be distributed to multiple network units. Part or all of the units can be selected according to actual needs to achieve the purpose of the solution of this embodiment.
- each functional unit in each embodiment of the present application may be integrated into one processing unit, each unit may exist separately physically, or two or more units may be integrated into one unit.
- “at least one” means one or more, and “multiple” means two or more.
- “And/or” describes the association relationship of associated objects, indicating that there may be three types of relationships, for example, A and/or B, which can mean: A exists alone, A and B exist simultaneously, and B exists alone, where A, B can be singular or plural.
- the character “/” generally indicates that the contextual objects are an “or” relationship; in the formulas of this application, the character “/” indicates that the contextual objects are a "division” Relationship.
- the functions described above are realized in the form of software function units and sold or used as independent products, they can be stored in a computer-readable storage medium.
- the technical solution of the present application is essentially or the part that contributes to the prior art or the part of the technical solution can be embodied in the form of a software product, and the computer software product is stored in a storage medium, including Several instructions are used to make a computer device (which may be a personal computer, a server, or a network device, etc.) execute all or part of the steps of the methods described in the various embodiments of the present application.
- the aforementioned storage media include: U disk, mobile hard disk, read-only memory (Read-Only Memory, ROM), random access memory (Random Access Memory, RAM), magnetic disk or optical disc and other media that can store program codes. .
Landscapes
- Engineering & Computer Science (AREA)
- Business, Economics & Management (AREA)
- Theoretical Computer Science (AREA)
- Finance (AREA)
- Accounting & Taxation (AREA)
- Physics & Mathematics (AREA)
- Strategic Management (AREA)
- Development Economics (AREA)
- General Physics & Mathematics (AREA)
- Data Mining & Analysis (AREA)
- General Business, Economics & Management (AREA)
- Marketing (AREA)
- Economics (AREA)
- Entrepreneurship & Innovation (AREA)
- General Engineering & Computer Science (AREA)
- Databases & Information Systems (AREA)
- Game Theory and Decision Science (AREA)
- Artificial Intelligence (AREA)
- Evolutionary Computation (AREA)
- Life Sciences & Earth Sciences (AREA)
- Software Systems (AREA)
- Computing Systems (AREA)
- Biophysics (AREA)
- Computational Linguistics (AREA)
- Health & Medical Sciences (AREA)
- General Health & Medical Sciences (AREA)
- Mathematical Physics (AREA)
- Biomedical Technology (AREA)
- Molecular Biology (AREA)
- Bioinformatics & Cheminformatics (AREA)
- Evolutionary Biology (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Bioinformatics & Computational Biology (AREA)
- Information Retrieval, Db Structures And Fs Structures Therefor (AREA)
Abstract
Description
样本 | 用户ID | 城市 | 身份 | 物品ID | 种类 | 商标 |
样本1 | U1 | LA | Student | L1 | T-shirt | B1 |
样本2 | U2 | NYC | Student | L1 | T-shirt | B1 |
样本3 | U1 | LA | Student | L2 | cell phone | B2 |
样本4 | U3 | London | manager | L3 | cell phone | B3 |
元素 | 文档 |
U1 | 样本1、样本3 |
U2 | 样本2 |
U3 | 样本4 |
LA | 样本1、样本3 |
NYC | 样本2 |
London | 样本4 |
Student | 样本1、样本2、样本3 |
L1 | 样本1、样本2 |
L2 | 样本3 |
L3 | 样本4 |
T-shirt | 样本1、样本2 |
cell phone | 样本3、样本4 |
B1 | 样本1、样本2 |
B2 | 样本3 |
B3 | 样本4 |
Claims (24)
- 一种推荐方法,其特征在于,包括:获取待预测数据;根据所述待预测数据和多个参考样本的相似度从所述多个参考样本中获取多个目标参考样本;每个所述参考样本和所述待预测数据均包括用户特征域数据和物品特征域数据,所述待预测数据的所述用户特征域数据用于指示目标用户特征,所述待预测数据的所述物品特征域数据用于指示目标物品特征,每个所述目标参考样本和所述待预测数据具有部分相同的用户特征域数据和/或物品特征域数据;根据所述多个目标参考样本与所述待预测数据获取所述待预测数据的目标特征信息;所述目标特征信息包括第一目标特征向量组和第二目标特征向量组,所述第一目标特征向量组为向量化后的所述待预测数据,所述第二目标特征向量组为对所述多个目标参考样本进行向量化后融合得到;以所述目标特征信息为输入通过深度神经网络DNN获取输出值;根据所述输出值确定是否向所述目标用户推荐所述目标物品。
- 根据权利要求1所述的方法,其特征在于,所述多个目标参考样本还包括标签数据;所述第二目标特征向量组为对所述多个目标参考样本进行向量化后融合得到,具体为:所述第二目标特征向量组为对所述多个目标参考样本的用户特征域数据,物品特征域数据以及标签数据进行向量化后融合得到。
- 根据权利要求1或2所述的方法,其特征在于,所述目标特征信息还包括第三目标特征向量组,所述第三目标特征向量组为对第一向量组中的目标特征向量进行两两交互得到,所述第一向量组包括所述第一目标特征向量组和所述第二目标特征向量组。
- 根据权利要求1-3中任一项所述的方法,其特征在于,所述融合包括:对所述第一目标特征向量组中的多个第一目标特征向量进行拼接,得到所述待预测数据的第二特征向量;对每个所述目标参考样本的多个第一特征向量进行拼接,得到每个所述目标参考样本的第二特征向量,每个所述目标参考样本的多个第一特征向量为对所述目标参考样本进行向量化得到;获取每个所述目标参考样本的第二特征向量与所述待预测数据的第二特征向量之间的相似度;根据每个所述目标参考样本的第二特征向量与所述待预测数据的第二特征向量之间的相似度,确定每个所述目标参考样本的权重;根据每个所述目标参考样本的权重,对所述多个目标参考样本在同一个特征域下的第一特征向量进行融合,得到所述第二目标特征向量组。
- 根据权利要求1-4中任一项所述的方法,其特征在于,根据所述待预测数据和多个参考样本的相似度从所述多个参考样本中获取多个目标参考样本之前,所述方法还包括:获取多个原始样本,其中,每个所述原始样本包括用户特征域数据和物品特征域数据;将所述待预测数据的多个所述用户特征域数据和多个所述物品特征域数据作为元素,对所述多个原始样本进行倒排索引,得到所述多个参考样本。
- 一种推荐模型训练方法,其特征在于,所述推荐模型包括特征信息提取网络和深度神经网络DNN,所述方法包括:获取多个训练样本,其中,每个所述训练样本包括用户特征域数据和物品特征域数据;根据第一训练样本和多个第二训练样本的相似度从所述多个第二训练样本中获取多个目标训练样本,其中,所述第一训练样本为所述多个训练样本中的一个,所述多个第二训练样本为所述多个训练样本除所述第一训练样本之外的部分或全部,所述第一训练样本的所述用户特征域数据用于指示第一参考用户特征,所述第一训练样本的所述物品特征域数据用于指示第一参考物品特征,所述第一训练样本和每个所述目标训练样本具有部分相同的用户特征域数据和/或物品特征域数据;将所述第一训练样本和所述多个目标训练样本输入到所述特征信息提取网络,得到所述第一训练样本的目标特征信息,其中,所述目标特征信息包括第四目标特征向量组和第五目标特征向量组,所述第四目标特征向量组为通过所述特征信息提取网络对所述第一训练样本进行向量化得到,所述第五目标特征向量组为通过所述特征信息提取网络对所述多个目标训练样本进行向量化后融合得到;将所述目标特征信息输入到所述深度神经网络DNN,得到输出值,所述输出值用于表征所述第一参考用户对所述第一参考物品进行操作的概率;根据所述输出值以及所述第一训练样本的标签数据进行所述推荐模型的训练,获得目标推荐模型。
- 根据权利要求6所述的方法,其特征在于,所述第五目标特征向量组为对所述多个目标训练样本进行向量化后融合得到,具体为:所述第五目标特征向量组为通过所述特征信息提取网络对所述多个目标训练样本的用户特征域数据、物品特征域数据以及标签数据进行向量化后融合得到。
- 根据权利要求6或7所述的方法,其特征在于,所述目标特征信息还包括第六目标特征向量组,所述第六目标特征向量组是通过所述特征信息提取网络对第二向量组中的目标特征向量进行两两交互得到,所述第二向量组包括所述第四目标特征向量组和所述第五目标特征向量组。
- 根据权利要求6-8中任一项所述的方法,其特征在于,所述融合包括:对所述第四目标特征向量组中的多个第四目标特征向量进行拼接,得到所述第一训练样本的第二特征向量;对每个所述目标训练样本的多个第一特征向量进行拼接,得到每个所述目标训练样本的第二特征向量,每个所述目标训练样本的多个第一特征向量为对所述目标训练样本进行向量化得到;获取每个所述目标训练样本的第二特征向量与所述第一训练样本的第二特征向量之间的相似度;根据每个所述目标训练样本的第二特征向量与所述第一训练样本的第二特征向量之间的相似度,确定每个所述目标训练样本的权重;根据每个所述目标训练样本的权重,对所述多个目标训练样本在同一个特征域下的第一特征向量进行融合,得到所述第五目标特征向量组。
- 根据权利要求6-9中任一项所述的方法,其特征在于,根据第一训练样本和多个第二训练样本的相似度从所述多个第二训练样本中获取多个目标训练样本之前,所述方法还包括:将所述第一训练样本的多个所述用户特征域数据和多个所述物品特征域数据作为元素,对所述多个训练样本进行倒排索引,得到所述多个第二训练样本。
- 一种推荐装置,其特征在于,包括:获取单元和处理单元;所述获取单元,用于获取待预测数据;所述处理单元,用于根据所述待预测数据和多个参考样本的相似度从所述多个参考样本中获取多个目标参考样本;每个所述参考样本和所述待预测数据均包括用户特征域数据和物品特征域数据,所述待预测数据的所述用户特征域数据用于指示目标用户特征,所述待预测数据的所述物品特征域数据用于指示目标物品特征,每个所述目标参考样本和所述待预测数据具有部分相同的用户特征域数据和/或物品特征域数据;根据所述多个目标参考样本与所述待预测数据获取所述待预测数据的目标特征信息;所述目标特征信息包括第一目标特征向量组和第二目标特征向量组,所述第一目标特征向量组为向量化后的所述待预测数据,所述第二目标特征向量组为对所述多个目标参考样本进行向量化后融合得到;以所述目标特征信息为输入通过深度神经网络DNN获取输出值;根据所述输出值确定是否向所述目标用户推荐所述目标物品。
- 根据权利要求11所述的装置,其特征在于,所述多个目标参考样本还包括标签数据;所述第二目标特征向量组为对所述多个目标参考样本进行向量化后融合得到,具体为:所述第二目标特征向量组为对所述多个目标参考样本的用户特征域数据,物品特征域数据以及标签数据进行向量化后融合得到。
- 根据权利要求12所述的装置,其特征在于,所述目标特征信息还包括第三目标特征向量组,所述第三目标特征向量组为对第一向量组中的目标特征向量进行两两交互得到,所述第一向量组包括所述第一目标特征向量组和所述第二目标特征向量组。
- 根据权利要求11-13中任一项所述的装置,其特征在于,在所述处理单元进行融合方面,所述处理单元,具体用于:对所述第一目标特征向量组中的多个第一目标特征向量进行拼接,得到所述待预测数据的第二特征向量;对每个所述目标参考样本的多个第一特征向量进行拼接,得到每个所述目标参考样本的第二特征向量,每个所述目标参考样本的多个第一特征向量为对所述目标参考样本进行向量 化得到;获取每个所述目标参考样本的第二特征向量与所述待预测数据的第二特征向量之间的相似度;根据每个所述目标参考样本的第二特征向量与所述待预测数据的第二特征向量之间的相似度,确定每个所述目标参考样本的权重;根据每个所述目标参考样本的权重,对所述多个目标参考样本在同一个特征域下的第一特征向量进行融合,得到所述第二目标特征向量组。
- 根据权利要求11-14中任一项所述的装置,其特征在于,在所述处理单元根据所述待预测数据和多个参考样本的相似度从所述多个参考样本中获取多个目标参考样本之前,所述处理单元,还用于获取多个原始样本,其中,每个所述原始样本包括用户特征域数据和物品特征域数据;将所述待预测数据的多个所述用户特征域数据和多个所述物品特征域数据作为元素,对所述多个原始样本进行倒排索引,得到所述多个参考样本。
- 一种推荐模型训练装置,其特征在于,所述推荐模型包括特征信息提取网络和深度神经网络DNN,所述装置包括:获取单元和处理单元;所述获取单元,用于获取多个训练样本,其中,每个所述训练样本包括用户特征域数据和物品特征域数据;所述处理单元,用于根据第一训练样本和多个第二训练样本的相似度从所述多个第二训练样本中获取多个目标训练样本,其中,所述第一训练样本为所述多个训练样本中的一个,所述多个第二训练样本为所述多个训练样本除所述第一训练样本之外的部分或全部,所述第一训练样本的所述用户特征域数据用于指示第一参考用户特征,所述第一训练样本的所述物品特征域数据用于指示第一参考物品特征,所述第一训练样本和每个所述目标训练样本具有部分相同的用户特征域数据和/或物品特征域数据;将所述第一训练样本和所述多个目标训练样本输入到所述特征信息提取网络,得到所述第一训练样本的目标特征信息,其中,所述目标特征信息包括第四目标特征向量组和第五目标特征向量组,所述第四目标特征向量组为通过所述特征信息提取网络对所述第一训练样本进行向量化得到,所述第五目标特征向量组为通过所述特征信息提取网络对所述多个目标训练样本进行向量化后融合得到;将所述目标特征信息输入到所述深度神经网络DNN,得到输出值,所述输出值用于表征所述第一参考用户对所述第一参考物品进行操作的概率;根据所述输出值以及所述第一训练样本的标签进行所述推荐模型的训练,获得目标推荐模型。
- 根据权利要求16所述的装置,其特征在于,所述第五目标特征向量组为对所述多个目标训练样本进行向量化后融合得到,具体为:所述第五目标特征向量组为通过所述特征信息提取网络对所述多个目标训练样本的用户特征域数据、物品特征域数据以及标签数据进行向量化后融合得到。
- 根据权利要求16或17所述的装置,其特征在于,所述目标特征信息还包括第六目标特征向量组,所述第六目标特征向量组是通过所述特征信息提取网络对第二向量组中的目标特征向量进行两两交互得到,所述第二向量组包括所述第四目标特征向量组和所述第五目标特征向量组。
- 根据权利要求16-18中任一项所述的装置,其特征在于,在所述处理单元进行融合方面,所述处理单元,具体用于:对所述第四目标特征向量组中的多个第四目标特征向量进行拼接,得到所述第一训练样本的第二特征向量;对每个所述目标训练样本的多个第一特征向量进行拼接,得到每个所述目标训练样本的第二特征向量,每个所述目标训练样本的多个第一特征向量为对所述目标训练样本进行向量化得到;获取每个所述目标训练样本的第二特征向量与所述第一训练样本的第二特征向量之间的相似度;根据每个所述目标训练样本的第二特征向量与所述第一训练样本的第二特征向量之间的相似度,确定每个所述目标训练样本的权重;根据每个所述目标训练样本的权重,对所述多个目标训练样本在同一个特征域下的第一特征向量进行融合,得到所述第五目标特征向量组。
- 根据权利要求16-19中任一项所述的装置,其特征在于,在所述处理单元根据第一训练样本和多个第二训练样本的相似度从所述多个第二训练样本中获取多个目标训练样本之前,所述处理单元,将所述第一训练样本的多个所述用户特征域数据和多个所述物品特征域数据作为元素,对所述多个训练样本进行倒排索引,得到所述多个第二训练样本。
- 一种电子设备,其特征在于,包括:存储器,用于存储程序;处理器,用于执行存储器存储的程序;当存储器存储的程序被执行时,处理器用于实现权利要求1-5或权利要求6-10中任一项所述的方法。
- 一种计算机可读存储介质,其特征在于,所述计算机可读存储介质存储用于设备执行的程序代码,所述程序代码包括用于实现权利要求1-5或权利要求6-10中任一项所述的方法。
- 一种计算程序产品,其特征在于,当所述计算程序产品在计算机上运行时,使得计算机实现权利要求1-5或权利要求6-10中任一项所述的方法。
- 一种芯片,其特征在于,所述芯片包括处理器与数据接口,所述处理器通过数据接口读取所述存储器上存储的指令,实现权利要求1-5或权利要求6-10中任一项所述的方法。
Priority Applications (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
EP22852095.3A EP4322031A1 (en) | 2021-07-31 | 2022-07-30 | Recommendation method, recommendation model training method, and related product |
US18/416,924 US20240202491A1 (en) | 2021-07-31 | 2024-01-19 | Recommendation method, method for training recommendation model, and related product |
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202110877429.9A CN113722583A (zh) | 2021-07-31 | 2021-07-31 | 推荐方法、推荐模型训练方法及相关产品 |
CN202110877429.9 | 2021-07-31 |
Related Child Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US18/416,924 Continuation US20240202491A1 (en) | 2021-07-31 | 2024-01-19 | Recommendation method, method for training recommendation model, and related product |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2023011382A1 true WO2023011382A1 (zh) | 2023-02-09 |
Family
ID=78674649
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/CN2022/109297 WO2023011382A1 (zh) | 2021-07-31 | 2022-07-30 | 推荐方法、推荐模型训练方法及相关产品 |
Country Status (4)
Country | Link |
---|---|
US (1) | US20240202491A1 (zh) |
EP (1) | EP4322031A1 (zh) |
CN (1) | CN113722583A (zh) |
WO (1) | WO2023011382A1 (zh) |
Cited By (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN116049411A (zh) * | 2023-03-31 | 2023-05-02 | 北京中关村科金技术有限公司 | 一种信息匹配方法、装置、设备和可读存储介质 |
CN116485560A (zh) * | 2023-06-21 | 2023-07-25 | 杭州大鱼网络科技有限公司 | 一种基于反馈机制的目标用户筛选方法及系统 |
CN117349458A (zh) * | 2023-12-05 | 2024-01-05 | 北京搜狐新媒体信息技术有限公司 | 一种多媒体推荐方法、装置、设备和存储介质 |
CN118035566A (zh) * | 2024-04-11 | 2024-05-14 | 中国科学技术大学 | 交互行为预测模型的训练方法、交互行为预测方法及装置 |
Families Citing this family (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN113722583A (zh) * | 2021-07-31 | 2021-11-30 | 华为技术有限公司 | 推荐方法、推荐模型训练方法及相关产品 |
CN114757700A (zh) * | 2022-04-12 | 2022-07-15 | 北京京东尚科信息技术有限公司 | 物品销量预测模型训练方法、物品销量预测方法及装置 |
CN115953215B (zh) * | 2022-12-01 | 2023-09-05 | 上海交通大学 | 一种基于时间和图结构的搜索式推荐方法 |
Citations (7)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN109299396A (zh) * | 2018-11-28 | 2019-02-01 | 东北师范大学 | 融合注意力模型的卷积神经网络协同过滤推荐方法及系统 |
CN111259263A (zh) * | 2020-01-15 | 2020-06-09 | 腾讯云计算(北京)有限责任公司 | 一种物品推荐方法、装置、计算机设备及存储介质 |
US20200272913A1 (en) * | 2018-11-30 | 2020-08-27 | Huawei Technologies Co., Ltd. | Recommendation Method and Apparatus |
CN111881363A (zh) * | 2020-06-23 | 2020-11-03 | 北京工业大学 | 一种基于图交互网络的推荐方法 |
CN112632403A (zh) * | 2020-12-24 | 2021-04-09 | 北京百度网讯科技有限公司 | 推荐模型的训练方法、推荐方法、装置、设备和介质 |
WO2021081962A1 (zh) * | 2019-10-31 | 2021-05-06 | 华为技术有限公司 | 推荐模型的训练方法、推荐方法、装置及计算机可读介质 |
CN113722583A (zh) * | 2021-07-31 | 2021-11-30 | 华为技术有限公司 | 推荐方法、推荐模型训练方法及相关产品 |
Family Cites Families (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN110232152B (zh) * | 2019-05-27 | 2021-03-23 | 腾讯科技(深圳)有限公司 | 内容推荐方法、装置、服务器以及存储介质 |
CN111008332B (zh) * | 2019-12-03 | 2023-04-07 | 深圳市雅阅科技有限公司 | 内容项推荐方法、装置、服务器以及存储介质 |
CN111008336A (zh) * | 2019-12-23 | 2020-04-14 | 腾讯科技(深圳)有限公司 | 一种内容推荐方法、装置、设备及可读存储介质 |
CN111461841B (zh) * | 2020-04-07 | 2023-04-07 | 腾讯云计算(北京)有限责任公司 | 物品推荐方法、装置、服务器及存储介质 |
CN111949886B (zh) * | 2020-08-28 | 2023-11-24 | 腾讯科技(深圳)有限公司 | 一种用于信息推荐的样本数据生成方法和相关装置 |
CN112287166B (zh) * | 2020-09-23 | 2023-03-07 | 山东师范大学 | 一种基于改进深度信念网络的电影推荐方法及系统 |
-
2021
- 2021-07-31 CN CN202110877429.9A patent/CN113722583A/zh active Pending
-
2022
- 2022-07-30 WO PCT/CN2022/109297 patent/WO2023011382A1/zh active Application Filing
- 2022-07-30 EP EP22852095.3A patent/EP4322031A1/en active Pending
-
2024
- 2024-01-19 US US18/416,924 patent/US20240202491A1/en active Pending
Patent Citations (7)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN109299396A (zh) * | 2018-11-28 | 2019-02-01 | 东北师范大学 | 融合注意力模型的卷积神经网络协同过滤推荐方法及系统 |
US20200272913A1 (en) * | 2018-11-30 | 2020-08-27 | Huawei Technologies Co., Ltd. | Recommendation Method and Apparatus |
WO2021081962A1 (zh) * | 2019-10-31 | 2021-05-06 | 华为技术有限公司 | 推荐模型的训练方法、推荐方法、装置及计算机可读介质 |
CN111259263A (zh) * | 2020-01-15 | 2020-06-09 | 腾讯云计算(北京)有限责任公司 | 一种物品推荐方法、装置、计算机设备及存储介质 |
CN111881363A (zh) * | 2020-06-23 | 2020-11-03 | 北京工业大学 | 一种基于图交互网络的推荐方法 |
CN112632403A (zh) * | 2020-12-24 | 2021-04-09 | 北京百度网讯科技有限公司 | 推荐模型的训练方法、推荐方法、装置、设备和介质 |
CN113722583A (zh) * | 2021-07-31 | 2021-11-30 | 华为技术有限公司 | 推荐方法、推荐模型训练方法及相关产品 |
Non-Patent Citations (1)
Title |
---|
XUE FENG, LIU KAI; WANG DONG; ZHANG HAOBO: "Personalized Recommendation Algorithm Based on Deep Neural Network and Weighted Implicit Feedback", PATTERN RECOGNITION AND ARTIFICIAL INTELLIGENCE, vol. 33, no. 4, 1 April 2020 (2020-04-01), pages 295 - 302, XP093033460, ISSN: 1003-6059, DOI: 10.16451/j.cnki.issn1003-6059.202004002 * |
Cited By (7)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN116049411A (zh) * | 2023-03-31 | 2023-05-02 | 北京中关村科金技术有限公司 | 一种信息匹配方法、装置、设备和可读存储介质 |
CN116049411B (zh) * | 2023-03-31 | 2023-06-27 | 北京中关村科金技术有限公司 | 一种信息匹配方法、装置、设备和可读存储介质 |
CN116485560A (zh) * | 2023-06-21 | 2023-07-25 | 杭州大鱼网络科技有限公司 | 一种基于反馈机制的目标用户筛选方法及系统 |
CN117349458A (zh) * | 2023-12-05 | 2024-01-05 | 北京搜狐新媒体信息技术有限公司 | 一种多媒体推荐方法、装置、设备和存储介质 |
CN117349458B (zh) * | 2023-12-05 | 2024-04-09 | 北京搜狐新媒体信息技术有限公司 | 一种多媒体推荐方法、装置、设备和存储介质 |
CN118035566A (zh) * | 2024-04-11 | 2024-05-14 | 中国科学技术大学 | 交互行为预测模型的训练方法、交互行为预测方法及装置 |
CN118035566B (zh) * | 2024-04-11 | 2024-08-23 | 中国科学技术大学 | 交互行为预测模型的训练方法、交互行为预测方法及装置 |
Also Published As
Publication number | Publication date |
---|---|
EP4322031A1 (en) | 2024-02-14 |
US20240202491A1 (en) | 2024-06-20 |
CN113722583A (zh) | 2021-11-30 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
WO2023011382A1 (zh) | 推荐方法、推荐模型训练方法及相关产品 | |
CN111784455B (zh) | 一种物品推荐方法及推荐设备 | |
EP3757905A1 (en) | Deep neural network training method and apparatus | |
CN111898031B (zh) | 一种获得用户画像的方法及装置 | |
CN113095415B (zh) | 一种基于多模态注意力机制的跨模态哈希方法及系统 | |
CN106447066A (zh) | 一种大数据的特征提取方法和装置 | |
CN109584006B (zh) | 一种基于深度匹配模型的跨平台商品匹配方法 | |
CN114332680A (zh) | 图像处理、视频搜索方法、装置、计算机设备和存储介质 | |
WO2024041483A1 (zh) | 一种推荐方法及相关装置 | |
CN111967924A (zh) | 商品推荐方法、商品推荐装置、计算机设备和介质 | |
CN111695024A (zh) | 对象评估值的预测方法及系统、推荐方法及系统 | |
WO2023185925A1 (zh) | 一种数据处理方法及相关装置 | |
CN110135769B (zh) | 货品属性填充方法及装置、存储介质及电子终端 | |
US20220172271A1 (en) | Method, device and system for recommending information, and storage medium | |
CN111368081A (zh) | 一种精选文本内容确定方法和系统 | |
CN113656589B (zh) | 对象属性确定方法、装置、计算机设备及存储介质 | |
CN112989182A (zh) | 信息处理方法、装置、信息处理设备及存储介质 | |
CN116910357A (zh) | 一种数据处理方法及相关装置 | |
EP4398128A1 (en) | Recommendation method and related device | |
CN116957128A (zh) | 业务指标预测方法、装置、设备和存储介质 | |
CN114298118B (zh) | 一种基于深度学习的数据处理方法、相关设备及存储介质 | |
CN116128581A (zh) | 物品推荐方法和系统 | |
CN114429384A (zh) | 基于电商平台的产品智能推荐方法及系统 | |
CN115408599A (zh) | 信息推荐方法、装置、电子设备及计算机可读存储介质 | |
CN112488355A (zh) | 基于图神经网络预测用户评级的方法和装置 |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 22852095 Country of ref document: EP Kind code of ref document: A1 |
|
WWE | Wipo information: entry into national phase |
Ref document number: 2022852095 Country of ref document: EP |
|
ENP | Entry into the national phase |
Ref document number: 2022852095 Country of ref document: EP Effective date: 20231106 |
|
WWE | Wipo information: entry into national phase |
Ref document number: 11202307996U Country of ref document: SG |
|
NENP | Non-entry into the national phase |
Ref country code: DE |