This disclosure relates generally to machine learning-based recommender systems and, more specifically, to an improved recommender system with less sample selection bias and better performance for small shops on an ecommerce platform.
A recommender system may be used to identify the users most likely to purchase an item for purposes of targeted advertising. It also may be used to recommend items to users. With the explosive growth of e-commerce in the past few decades, there are more product offerings than consumers can process, and recommender systems have become imperative to overcome this overload problem. The goal of recommender systems is to take known user feedback and predict unknown user feedback on items, which can be helpful in targeting advertisements for a product or for providing a list of suggested items to a user. The feedback, often referred to as the interaction, can be implicit (e.g., purchased/likely to purchase) or explicit (e.g., a rating between 1 and 5).
Many recommender systems use machine learning models to make predictions. For example, neural collaborative filtering (NCF) systems apply neural network encoders to user and item data to generate representations of the user and item data, and then use the user and item representations to predict user ratings for the items.
The machine learning models used to make predictions for an ecommerce platform are typically trained based on user-items interactions in the ecommerce platform. There is significantly more training data for shops with large sales volumes on the platform than for small shops. As a result, these models are better at making predictions for large shops than for small shops. This is what is known as sample selection bias. In fact, the performance of such models with respect to a new item for a small shop can be poor. Therefore, there is demand for a recommender system with good prediction performance for both small and large shops on an ecommerce platform.
The present disclosure relates to improving recommendations for small shops on an ecommerce platform while maintaining accuracy for larger shops. The improvement is achieved by retraining a machine-learning recommendation model to reduce sample selection bias using a meta-learning process. The retraining process comprises identifying a sample subset of shops on the ecommerce platform, and then creating shop-specific versions of the recommendation model for each of the shops in the subset. Each shop-specific model is created by optimizing the baseline model to predict user-item interactions for a shop using a first training dataset for the shop. Each of the shop-specific models is then tested using a second training dataset for the shop. A loss is calculated for each shop-specific model based on the model's predicted user-item interactions for the applicable second training dataset and the actual user-item interactions in the second training dataset. A global loss is calculated based on each of the shop-specific losses, and the parameters of the baseline model are updated to minimize the global loss.
The global model includes small and large-shop weight parameters that are applied to user-item interaction scores, depending on whether the input item is from a small shop or a large shop. The weight parameters are learned during the retraining process. In learning the different weight parameters for small and large shops, the system learns the difference between recommendations for small shops and recommendations for large shops. This mitigates the model bias towards large shops, and results in improved recommendation performance for both large and small shops.
In one embodiment, the retraining process comprises the following steps:
The present disclosure relates to a machine learning-based recommender system and method for predicting user-item interactions on an ecommerce platform that includes shops with different sales volumes. Specifically, the disclosure relates to an improved recommender system with better predictions for small shops while still maintaining prediction accuracy for larger shops. The improvement is achieved by retraining the recommender system to reduce sample selection bias using a meta-learning process. The methods disclosed herein are performed by a computer system (“the system”).
We define small/large shops based on the number of sales on the applicable ecommerce platform during a period of time relative to a threshold. Specifically, the terms “small shop” and “smaller shop” refers herein to shops with a sales volume below a certain threshold. Conversely, the terms “large shop,” “larger shop” and “largest shops” refers to shops with a sales volume above a certain threshold. The invention is not dependent or limited to any particular threshold or range of thresholds. The threshold that defines small and large shops may depend on the particular ecommerce platform at issue and may be different for each ecommerce platform. However, as example, the threshold may be the median shop sales volume for all shops on the ecommerce platform during a period of time, such as the past 12 months.
Both a training phase and a prediction phase are described below for the recommender system.
The system retrains the baseline model to reduce the sample selection bias (step 120). Retraining the model involves identifying a sample batch of shops on the ecommerce platform (step 120a). The sample batch is a subset of shops on the platform, and it includes shops in the small size category and shops in the large size category.
The system obtains two training datasets, referred to herein as a first training dataset and a second training dataset, for each of the shops in the sample batch (step 120b). Each of the first and second training datasets for a shop includes: (1) item data for items in the shop, and (2) user data, including user-item interaction data. Examples of user-item interaction data include user ratings or purchase history (i.e., whether or not a user has purchased an item).
The system performs the retraining method depicted in
1.1 Retraining Method
1.1.1 Creating Shop-Specific Versions of the Model
The system applies a global model (200) to user and item data in the first training dataset (210) for a shop. The global model generates a preliminary user-item interaction score (212) for each user-item pair in the shop's first training dataset. The global model is the machine learning model for all shops on the ecommerce platform as opposed to a shop-specific version of the model. During the first iteration of the retraining method, the global model is the baseline model. In subsequent iterations, it is an updated version of the baseline model.
The system determines if the shop is in a small-size category or a large-size category. In response to the shop being in the small-size category, the system applies a global small-shop weight parameter (214) to the preliminary user-item interaction scores to obtain final user-item interaction scores for the first dataset. In response to the shop being in the large-size category, the system applies a global large-shop weight parameter (213) to the preliminary user-item interaction scores to obtain final user-item interaction scores for the first dataset. The system generates user-item predictions (212) using the final user-item interaction scores.
The system calculates a first shop-specific loss (215) for the shop based on the predicted user and item interactions and the actual user and item interactions in the first dataset for the shop. In one embodiment, mean square error is used as the loss function. The system calculates a shop-specific model update (220) to minimize the loss. This comprises calculating shop-specific parameter adjustments to the model. The system then creates a shop-specific version of the model (225) by applying the first shop-specific update to the model. In certain embodiments, the system also calculates an adjustment to the small or large weight parameter (whichever is applicable based on the size of the shop) to minimize the shop-specific loss. In other words, in certain embodiments, the system also creates a local, shop-specific weight parameter (small or large, whichever is applicable) that will be used in making user-item predictions for the shop's second training dataset.
In one embodiment, the first training dataset is the same size for each of the shops. This puts small shops on par with larger shops for this phase of the training. In other words, using the same size dataset for each shop in this phase prevents overfitting to larger shops that have more sales data than smaller shops and reduces the sample selection bias inherent in the baseline model.
As discussed below, the next part of the retraining method comprises applying the shop-specific models to the second training datasets and using the corresponding losses to make a global parameter adjustment to the global model.
1.1.2 Using the Shop-Specific Models to Identify Aa Global Parameter Adjustment
For each shop in the sample batch, the system applies the shop-specific version of the model (225) to user and item data in the second training dataset (230) for the shop to obtain preliminary user and item interaction scores (235) for the second dataset. In one embodiment, the size of the second dataset corresponds to the sales volume for the shop on the ecommerce platform, up to a maximum number (e.g., 100 k transactions). This helps to maintain the prediction accuracy of the model for larger shops.
If the shop is in the small size category, the system applies the small-shop weight parameter (250) to the preliminary user-item interaction scores to obtain final user-item interaction scores for the second training dataset. If the shop is in the large size category, then the system applies the large-shop weight parameter (240) to the preliminary user-item interaction scores to obtain final user-item interaction scores for the second training dataset. In embodiments where a shop-specific weight parameter adjustment was calculated after the first shop-specific loss, the weight parameter applied is a shop-specific (local) weight parameter.
The system calculates a second shop-specific loss (215) for the shop based on the predicted user and item interactions and the actual user and item interactions in the second dataset for the shop. In
The system calculates a global loss (260) based on all the second shop-specific losses. In one embodiment, the global loss is calculated as an average of the second shop specific losses. The system then calculates a global parameter adjustment (270) for the model to reduce the global loss. The system creates an updated global model by adjusting the parameters of the model using the global parameter adjustment. In one embodiment, the shop-specific and global parameter adjustments are calculated using gradient descent, and the shop-specific and global parameter adjustments are gradient steps.
The system also makes adjustments to the global small and large weight parameters to minimize the global loss. In learning the different weight parameters for small and large shops, the system learns the difference between recommendations for small shops and recommendations for large shops. This mitigates the model bias towards large shops, and results in improved recommendation performance for both large and small shops.
The system repeats the retraining method for a number of iterations, wherein the updated global model in the previous iteration becomes the global model in the next iteration, and the updated global small and large weight parameters are the global small and large weight parameters in the next iteration.
1.1.3 Mathematical Expression of Retraining Process
Below is a mathematical expression of the retraining method, according to one embodiment.
from
hop;
do
Sp and
Qp from
p;
(
Sp; θp) ;
is small ∇θ
(
Qp; θp);
is large ∇θ
(
Qp; θp);
In the algorithm above:
In one embodiment, the loss function is defined as (Dp, θ)=loss (y, ŷ), where y is the actual purchase label (0 for not purchase, 1 for purchase) and ŷ=gθ(fu,fi) is the predicted label, where fu,fi are user feature and item feature, which can be trainable one-hot embeddings or pretrained representations.
In a prediction phase, the retrained model can be used either to recommend users to shops for targeted advertisements or to recommend items to users on the ecommerce platform.
In
In one embodiment, the input user data includes user-item interaction data. It may also include “side information” about a user (e.g., user demographics, such as user age, location, etc.). In one embodiment, the item data includes item “side information” which is information about the item (e.g., product category and subcategories).
The methods described with respect to
As will be understood by those familiar with the art, the invention may be embodied in other specific forms without departing from the spirit or essential characteristics thereof. Accordingly, the above disclosure is intended to be illustrative, but not limiting, of the scope of a claimed invention.
This application claims the benefit of U.S. Provisional Application No. 63/221,872 filed on Jul. 14, 2021, and titled “Improving Cold-Start Item Advertisement for Small Businesses,” the contents of which are incorporated by reference herein as if fully disclosed herein.
Number | Name | Date | Kind |
---|---|---|---|
7987188 | Neylon | Jul 2011 | B2 |
8386336 | Fox | Feb 2013 | B1 |
8417713 | Blair-Goldensohn | Apr 2013 | B1 |
8458054 | Thakur | Jun 2013 | B1 |
10354184 | Vitaladevuni et al. | Jul 2019 | B1 |
10614381 | Hoffman et al. | Apr 2020 | B2 |
10698967 | Shen et al. | Jun 2020 | B2 |
10769524 | Natesh | Sep 2020 | B1 |
11004135 | Sandler et al. | May 2021 | B1 |
11361365 | Greenwald | Jun 2022 | B2 |
11651037 | Shi et al. | May 2023 | B2 |
11669759 | Horowitz et al. | Jun 2023 | B2 |
20010021914 | Jacobi | Sep 2001 | A1 |
20050162670 | Shuler | Jul 2005 | A1 |
20060155684 | Liu | Jul 2006 | A1 |
20070046675 | Iguchi | Mar 2007 | A1 |
20070087756 | Hoffberg | Apr 2007 | A1 |
20080270363 | Hunt | Oct 2008 | A1 |
20080294996 | Hunt | Nov 2008 | A1 |
20080319829 | Hunt | Dec 2008 | A1 |
20090006156 | Hunt | Jan 2009 | A1 |
20090018996 | Hunt | Jan 2009 | A1 |
20090110089 | Green | Apr 2009 | A1 |
20090125371 | Neylon | May 2009 | A1 |
20090281923 | Selinger et al. | Nov 2009 | A1 |
20100114933 | Murdock | May 2010 | A1 |
20100268661 | Levy | Oct 2010 | A1 |
20140104495 | Preston et al. | Apr 2014 | A1 |
20140195931 | Kwon | Jul 2014 | A1 |
20140321761 | Wang | Oct 2014 | A1 |
20140330637 | Moran | Nov 2014 | A1 |
20140344013 | Karty | Nov 2014 | A1 |
20140351079 | Dong | Nov 2014 | A1 |
20150112790 | Wolinsky | Apr 2015 | A1 |
20150154229 | An et al. | Jun 2015 | A1 |
20150154508 | Chen | Jun 2015 | A1 |
20150332374 | Fano | Nov 2015 | A1 |
20150379732 | Sayre, III | Dec 2015 | A1 |
20160155173 | Isaacson | Jun 2016 | A1 |
20160180248 | Regan | Jun 2016 | A1 |
20160292148 | Aley | Oct 2016 | A1 |
20170185894 | Volkovs et al. | Jun 2017 | A1 |
20170193011 | Kale | Jul 2017 | A1 |
20170193997 | Chen | Jul 2017 | A1 |
20180040064 | Grigg | Feb 2018 | A1 |
20180158078 | Hsieh | Jun 2018 | A1 |
20180204111 | Zadeh | Jul 2018 | A1 |
20180276710 | Tietzen | Sep 2018 | A1 |
20180308112 | Prentice | Oct 2018 | A1 |
20190019016 | Ikeda | Jan 2019 | A1 |
20190034875 | Bryan et al. | Jan 2019 | A1 |
20190244270 | Kim | Aug 2019 | A1 |
20200004835 | Ramanath et al. | Jan 2020 | A1 |
20200004886 | Ramanath et al. | Jan 2020 | A1 |
20200005134 | Ramanath et al. | Jan 2020 | A1 |
20200005149 | Ramanath et al. | Jan 2020 | A1 |
20200005364 | Aznaurashvili | Jan 2020 | A1 |
20200175022 | Nowozin | Jun 2020 | A1 |
20200211065 | Govindarajalu | Jul 2020 | A1 |
20200380027 | Aggarwal et al. | Dec 2020 | A1 |
20210004437 | Zhang et al. | Jan 2021 | A1 |
20210012150 | Liu et al. | Jan 2021 | A1 |
20210073612 | Vahdat | Mar 2021 | A1 |
20210081462 | Lu et al. | Mar 2021 | A1 |
20210097400 | Lee | Apr 2021 | A1 |
20210110306 | Krishnan et al. | Apr 2021 | A1 |
20210117839 | Kulkarni et al. | Apr 2021 | A1 |
20210133846 | Xu et al. | May 2021 | A1 |
20210150337 | Raziperchikolaei | May 2021 | A1 |
20210191990 | Shi et al. | Jun 2021 | A1 |
20210350393 | Dagley et al. | Nov 2021 | A1 |
20210382935 | Huang | Dec 2021 | A1 |
20210383254 | Renders et al. | Dec 2021 | A1 |
20210397892 | Huang | Dec 2021 | A1 |
20220114643 | Raziperchikolaei | Apr 2022 | A1 |
20220155940 | Olbrich et al. | May 2022 | A1 |
20220207073 | Sohail et al. | Jun 2022 | A1 |
20220277741 | Chaudhary et al. | Sep 2022 | A1 |
20220300804 | Guan et al. | Sep 2022 | A1 |
20220414531 | Ong et al. | Dec 2022 | A1 |
20230036394 | Shi | Feb 2023 | A1 |
20230036964 | Shi | Feb 2023 | A1 |
20230055699 | Raziperchikolaei | Feb 2023 | A1 |
Number | Date | Country |
---|---|---|
110019652 | Jul 2019 | CN |
110309331 | Oct 2019 | CN |
Entry |
---|
Gharibshah, Zhabiz, et al. “Deep learning for user interest and response prediction in online display advertising.” Data Science and Engineering 5.1 (2020): 12-26. (Year: 2020). |
Bohanec, Marko, Mirjana Kljajic Borstnar, and Marko Robnik-Šikonja. “Explaining machine learning models in sales predictions.” Expert Systems with Applications 71 (2017) (Year: 2017). |
Agarwal, Pankaj et al., “Personalizing Similar Product Recommendations in Fashion E-commerce”, Jun. 29, 2018, 5 pages. |
Bhaskar, Karthik Raja Kalaiselvi et al., “Implicit Feedback Deep Collaborative Filtering Product Recommendation System”, Sep. 8, 2020, 10 pages. |
Bronstein et al., “Data Fusion through Cross-modality Metric Learning using Similarity-Sensitive Hashing”, 2010, pp. 1-8. |
Cao et al., “Collective Deep Quantization for Efficient Cross-Modal Retrieval”, Proceedings of the Thirty-First AAAI Conference on Artificial Intelligence, 2017, pp. 3974-3980. |
Cao et al., “Deep Visual-Semantic Hashing for Cross-Modal Retrieval”, KDD '16: Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, Aug. 2016, pp. 1445-1454. |
Cao et al., “Correlation Hashing Network for Efficient Cross-Modal Retrieval”, 2016, pp. 1-12. |
Chen, Jingyuan, et al. “Attentive Collaborative Filtering: Multimedia Recommendation with Item-and Component-Level Attention”, Proceedings of the 40th International ACM SIGIR Conference on Research and Development in Information Retrieval, 2017, 10 pages. |
Deng et al., “DeepCF: A Unified Framework of Representation Learning and Matching Function Learning in Recommender System”, 2019, pp. 1-9. |
Ding et al., “Collective Matrix Factorization Hashing for Multimodal Data”, 2014, pp. 4321-4328. |
Dong et al., “A Hybrid Collaborative Filtering Model with Deep Structure for Recommender Systems”, Proceedings of the Thirty-First AAAI Conference on Artificial Intelligence, 2017, pp. 1309-1315. |
Gong et al., “Learning Binary Codes for High-Dimensional Data Using Bilinear Projections”, 2013, pp. 484-491. |
Guo et al., “DeepFM: A Factorization-Machine based Neural Network for CTR Prediction”, Proceedings of the Twenty- Sixth International Joint Conference on Artificial Intelligence, 2017, pp. 1725-1731. |
Kanagala, Mukhul “Product Recommendation System Using Machine Learning Techniques”, California State University San Marcos, Dec. 10, 2020, pp. 1-32. |
He et al., “Neural Factorization Machines for Sparse Predictive Analytics”, SIGIR '17, Aug. 7-11, 2017, pp. 355-364. |
He et al., “Outer Product-based Neural Collaborative Filtering”, Proceedings of the Twenty-Seventh International Joint Conference on Artificial Intelligence 2018, pp. 2227-2233. |
Jiang et al., “Deep Cross-Modal Hashing”, 2017, pp. 3232-3240. |
Kumar et al., “Learning Hash Functions for Cross-View Similarity Search”, 2011, pp. 1-6. |
Li et al., “Deep Binary Reconstruction for Cross-modal Hashing”, MM '17, Oct. 23-27, 2017, pp. 1-8. |
Li et al., “Deep Collaborative Filtering via Marginalized Denoising Auto-encoder”, CIKM '15, Oct. 19-23, 2015, pp. 811-820. |
Li et al., “Coupled Cycle-GAN: Unsupervised Hashing Network for Cross-Modal Retrieval”, Thirty-Third AAAI Conference on Artificial Intelligence, 2019, pp. 176-183. |
Li et al., “Deep Heterogeneous Autoencoders for Collaborative Filtering”, 2018, pp. 1-6. |
Li et al., “Self-Supervised Adversarial Hashing Networks for Cross-Modal Retrieval”, 2018, pp. 4242-4251. |
Lian et al., “xDeepFM: Combining Explicit and Implicit Feature Interactions for Recommender Systems”, KDD 2018, Aug. 19-23, 2018, pp. 1-10. |
Liu et al., “Recommender Systems with Heterogeneous Side Information”, WWW '19, May 13-17, 2019, pp. 1-7. |
Long et al., “Composite Correlation Quantization for Efficient Multimodal Retrieval”, SIGIR '16, Jul. 17-21, 2016, pp. 1-11. |
Raziperchikolaei, Ramin, et al. “Neural Representations in Hybrid Recommender Systems: Prediction versus Regularization”, Proceedings of the 44th International ACM SIGIR Conference on Research and Development in Information Retrieval, 2021, pp. 1743-1747. |
Ricci et al., “Recommender Systems Handbook”, 2011, 845 pages. |
Sedhain et al., “AutoRec: Autoencoders Meet Collaborative Filtering”, WWWW 2015 Companion, May 18-22, 2015, pp. 1-2. |
Strub et al., “Hybrid Recommender System based on Autoencoders”, Workshop on Deep Learning for Recommender Systems, Sep. 2016, pp. 1-7. |
Su et al., “Deep Joint-Semantics Reconstructing Hashing for Large-Scale Unsupervised Cross-Modal Retrieval”, 2019, pp. 3027-3035. |
Takács, Gábor, et al. “Matrix Factorization and Neighbor Based Algorithms for the Netflix Prize Problem”, Proceedings of the 2008 ACM Conference on Recommender Systems, 2008, pp. 267-274. |
Wan et al., “Discriminative Latent Semantic Regression for Cross-Modal Hashing of Multimedia Retrieval”, 2018 IEEE Fourth International Conference on Multimedia Big Data (BigMM), Oct. 21, 2018, pp. 1-7. |
Wang et al., “Collaborative Deep Learning for Recommender Systems”, KDD '15, Aug. 10-13, 2015, pp. 1235-1244. |
Wang et al., “Effective Multi-Modal Retrieval based on Stacked Auto-Encoders”, Proceedings of the VLDB Endowment, Sep. 1-5, 2014, pp. 649-660. |
Wang, Huiwei et al., “ML2E: Meta-Learning Embedding Ensemble for Cold-Start Recommendation”, IEEE Access, Sep. 2020, pp. 165757-165768. |
Wu et al., “Quantized Correlation Hashing for Fast Cross-Modal Search”, Proceedings of the Twenty-Fourth International Joint Conference on Artificial Intelligence, 2015, pp. 3946-3952. |
Wu et al., “Unsupervised Deep Hashing via Binary Latent Factor Models for Large-scale Cross-modal Retrieval”, Proceedings of the Twenty-Seventh International Joint Conference on Artificial Intelligence, 2018, p. 2854-2860. |
Yang et al., “Shared Predictive Cross-Modal Deep Quantization”, IEEE Transactions on Neural Networks and Learning Systems, vol. 29, No. 11, Nov. 2018, pp. 5292-5303. |
Yu, Runsheng et al., “Personalized Adaptive Meta Learning for Cold-Start User Preference Prediction”, 35th AAAI Conference on Artificial Intelligence, Feb. 2021, pp. 10772-10780. |
Zhang et al., “Collaborative Quantization for Cross-Modal Similarity Search”, 2019, pp. 1-10. |
Zhang et al., “Large-Scale Multimodal Hashing with Semantic Correlation Maximization”, Association for the Advancement of Artificial Intelligence, 2014, pp. 1-7. |
He, Xiangnan, et al. “NAIS: Neural Attentive Item Similarity Model for Recommendation”, IEEE Transactions on Knowledge and Data Engineering, 2018, 13 pages. |
Koren, “Factorization Meets the Neighborhood: a Multifaceted Collaborative Filtering Model”, KDD 2008, Aug. 24-27, 2008, pp. 426-434. |
Koren, Yehuda, et al. “Matrix Factorization Techniques for Recommender Systems”, Computer, Published by IEEE Computer Society, 2009, pp. 42-49. |
Nahta, Ravi et al., “Embedding metadata using deep collaborative filtering to address the cold start problem for the rating prediction task”, Multimedia Tools and Applications, vol. 80, No. 12, Feb. 18, 2021, pp. 18553-18581. |
Salakhutdinov, Russ, et al. “Probabilistic Matrix Factorization”, Advances in Neural Information Processing Systems, 2007, pp. 1-8. |
Xue, Hong-Jian, et al. “Deep Matrix Factorization Models for Recommender Systems”, Proceedings of the Twenty-Sixth International Joint Conference on Artificial Intelligence, vol. 17, 2017, pp. 3203-3209. |
Zhang, Yongfeng, et al. “Joint Representation Learning for Top-N Recommendation with Heterogeneous Information Sources”, Proceedings of the 2017 ACM Conference on Information and Knowledge Management, 2017, pp. 1-10. |
Hooda, Rahul et al., “Social Commerce Hybrid Product Recommender”, International Journal of Computer Applications, vol. 100, No. 12, Aug. 2014, pp. 43-49. |
Zhao, Tong “Learning to Search and Recommend From Users Implicit Feedback”, Aug. 2018, 209 pages. |
Gharibshah, Zhabiz et al., “User Response Prediction in Online Advertising”, ACM Comput. Surv., vol. 37, No. 4, Article 111, Aug. 2021, pp. 1-49. |
Ma, Yifei et al., “Temporal-Contextual Recommendation in Real-Time”, KDD '20, Aug. 23-27, 2020, pp. 2291-2299. |
Bianchi, Federico et al., “Fantastic Embeddings and How to Align Them: Zero-Shot Inference in a Multi-Shop Scenario”, SIGIR eCom '20, Jul. 30, 2020, pp. 1-11. |
Antoniou, Antreas et al., “How to Train Your MAML”, ICLR 2019. |
Bansal, Trapit et al., “Learning to Few-Shot Leam Across Diverse Natural Language Classification Tasks”, Proceedings of the 28th International Conference on Computational Linguistics, Dec. 2020, pp. 5108-5123. |
Barkan, Oren et al. “CB2CF: A Neural Multiview Content-to-Collaborative Filtering Model for Completely Cold Item Recommendations”, Proceedings of the 13th ACM Conference on Recommender Systems, 2019, pp. 1-9. |
Blei, David M. et al. “Latent Dirichlet Allocation”, Journal of Machine Learning Research, 2003, pp. 993-1022. |
Cai, Qi et al., “Memory Matching Networks for One-Shot Image Recognition”, Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, 2018, pp. 4080-4088. |
Chen, Minmin et al. “Marginalized Denoising Autoencoders for Domain Adaptation”, Proceedings of the 29th International Conference on Machine Learning, 2012. |
Chen, Zhihong et al. “ESAM: Discriminative Domain Adaptation with Non-Displayed Items to Improve Long-Tail Performance”, Proceedings of the 43rd International ACM SIGIR Conference on Research and Development in Information Retrieval, 2020, pp. 579-588. |
Cheng et al., “Wide & Deep Learning for Recommender Systems”, DLRS '16, Sep. 15, 2016, pp. 1-4. |
Chopra, Sumit et al., “Learning a Similarity Metric Discriminatively, with Application to Face Verification”, 2005 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR'05), vol. 1. IEEE, 2005. |
Dong, Xin et al. “A Hybrid Collaborative Filtering Model with Deep Structure for Recommender Systems”, Proceedings of the Thirty-First AAAI Conference on Artificial Intelligence, 2017, vol. 31, No. 1, pp. 1309-1315. |
Dong, Manqing et al., “MAMO: Memory-Augmented Meta-Optimization for Cold-Start Recommendation”, Proceedings of the 26th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining, 2020. |
Du, Zhengxiao et al. “Sequential Scenario-Specific Meta Learner for Online Recommendation”, Proceedings of the 25th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining, 2019, pp. 2895-2904. |
Finn, Chelsea et al. “Model-Agnostic Meta-Learning for Fast Adaptation of Deep Networks”, Proceedings of the 34th International Conference on Machine Learning, vol. 70, 2017, pp. 1126-1135. |
Gao, Chen et al. “Cross-domain Recommendation Without Sharing User-relevant Data”, The World Wide Web Conference, 2019, pp. 491-502. |
Gopalan, Prem et al., “Scalable Recommendation with Hierarchical Poisson Factorization”, UAI, 2015. |
He, Xiangnan et al. “Neural Collaborative Filtering”, Proceedings of the 26th International Conference on World Wide Web, 2017. |
Krishnan, Adit et al., “An Adversarial Approach to Improve Long-Tail Performance in Neural Collaborative Filtering”, Proceedings of the 27th ACM International Conference on Information and Knowledge Management, 2018, pp. 1491-1494. |
Lee, Hoyeop et al. “MeLU: Meta-Learned User Preference Estimator for Cold-Start Recommendation”, Proceedings of the 25th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining, 2019, pp. 1073-1082. |
Li, Sheng et al. “Deep Collaborative Filtering via Marginalized Denoising Auto-encoder”, Proceedings of the 24th ACM International Conference on Information and Knowledge Management, 2015, pp. 811-820. |
Li, Tianyu et al. “Deep Heterogeneous Autoencoders for Collaborative Filtering”, 2018 IEEE International Conference on Data Mining (ICDM), IEEE, 2018. |
Linden, Greg et al., “Amazon.com Recommendations: Item-to-Item Collaborative Filtering”, IEEE Internet Computing, 2003, pp. 76-80. |
Liu, Yudan et al. “Real-time Attention Based Look-alike Model for Recommender System”, Proceedings of the 25th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining, 2019. |
Luo, Mi et al., “Metaselector: Meta-Learning for Recommendation with User-Level Adaptive Model Selection”, Proceedings of The Web Conference, 2020, pp. 2507-2513. |
Mooney, Raymond J., et al. “Content-Based Book Recommending Using Learning for Text Categorization”, Proceedings of the Fifth ACM conference on Digital Libraries, 2000. |
Pan, Feiyang et al., “Warm Up Cold start Advertisements: Improving CTR Predictions via Learning to Learn ID Embeddings”, Proceedings of the 42nd International ACM SIGIR Conference on Research and Development in Information Retrieval, 2019. |
Raziperchikolaei, Ramin et al., “Shared Neural Item Representations for Completely Cold Start Problem”, Fifteenth ACM Conference on Recommender Systems, 2021, pp. 422-431. |
Shi, Shaoyun et al. “Attention-based Adaptive Model to Unify Warm and Cold Starts Recommendation”, Proceedings of the 27th ACM International Conference on Information and Knowledge Management, 2018, pp. 127-136. |
Slack, Dylan et al., “Fairness Warnings and Fair-MAML: Learning Fairly with Minimal Data”, Proceedings of the 2020 Conference on Fairness, Accountability, and Transparency, 2020, pp. 200-209. |
Van den Oord, Aaron et al. “Deep content-based music recommendation”, Advances in Neural Information Processing Systems 26 (2013), pp. 1-9. |
Vartak, Manasi et al. “A Meta-Learning Perspective on Cold-Start Recommendations for Items”, Advances in Neural Information Processing Systems, 2017. |
Vilalta, Ricardo et al., “A Perspective View and Survey of Meta-Learning”, Artificial Intelligence Review, Sep. 2001, pp. 77-95. |
Vincent, Pascal et al. “Stacked Denoising Autoencoders: Learning Useful Representations in a Deep Network with a Local Denoising Criterion”, Journal of Machine Learning Research, 2010, pp. 3371-3408. |
Volkovs, Maksims et al. “Dropoutnet: Addressing Cold Start in Recommender Systems”, Advances in Neural Information Processing Systems, 2017, pp. 1-10. |
Wang, Hao et al. “Collaborative Deep Learning for Recommender Systems”, Proceedings of the 21th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, 2015, pp. 1235-1244. |
Wang, Chong et al. “Collaborative Topic Modeling for Recommending Scientific Articles”, Proceedings of the 17th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, 2011. |
Yin, Wenpeng “Meta-learning for Few-shot Natural Language Processing: A Survey”, Jul. 2020. |
Yuan, Bowen et al. “Improving Ad Click Prediction by Considering Non-displayed Events”, Proceedings of the 28th ACM International Conference on Information and Knowledge Management, 2019. |
Zhang, Shuai et al., “Autosvd++: An Efficient Hybrid Collaborative Filtering Model via Contractive Auto-encoders”, Proceedings of the 40th International ACM SIGIR Conference on Research and Development in Information Retrieval, 2017, pp. 957-960. |
Zhang, Yin et al. “A Model of Two Tales: Dual Transfer Learning Framework for Improved Long-tail Item Recommendation”, Proceedings of the Web Conference 2021, pp. 2220-2231. |
Zhang, Yang et al., “How to Retrain Recommender System? A Sequential Meta-Learning Method”, Proceedings of the 3rd International ACM SIGIR Conference on Research and Development in Information Retrieval, 2020, pp. 1479-1488. |
Zhu, Yongchun et al. “Learning to Warm Up Cold Item Embeddings for Cold-Start Recommendation with Meta Scaling and Shifting Networks”, Proceedings of the 44th International ACM SIGIR Conference on Research and Development in Information Retrieval, 2021, pp. 1167-1176. |
Number | Date | Country | |
---|---|---|---|
20230036394 A1 | Feb 2023 | US |
Number | Date | Country | |
---|---|---|---|
63221872 | Jul 2021 | US |