skip to main content
10.1145/3397271.3401167acmconferencesArticle/Chapter ViewAbstractPublication PagesirConference Proceedingsconference-collections
research-article

How to Retrain Recommender System?: A Sequential Meta-Learning Method

Authors Info & Claims
Published:25 July 2020Publication History

ABSTRACT

Practical recommender systems need be periodically retrained to refresh the model with new interaction data. To pursue high model fidelity, it is usually desirable to retrain the model on both historical and new data, since it can account for both long-term and short-term user preference. However, a full model retraining could be very time-consuming and memory-costly, especially when the scale of historical data is large. In this work, we study the model retraining mechanism for recommender systems, a topic of high practical values but has been relatively little explored in the research community.

Our first belief is that retraining the model on historical data is unnecessary, since the model has been trained on it before. Nevertheless, normal training on new data only may easily cause overfitting and forgetting issues, since the new data is of a smaller scale and contains fewer information on long-term user preference. To address this dilemma, we propose a new training method, aiming to abandon the historical data during retraining through learning to transfer the past training experience.Specifically, we design a neural network-based transfer component, which transforms the old model to a new model that is tailored for future recommendations. To learn the transfer component well, we optimize the "future performance'' -- i.e., the recommendation accuracy evaluated in the next time period. Our Sequential Meta-Learning(SML) method offers a general training paradigm that is applicable to any differentiable model. We demonstrate SML on matrix factorization and conduct experiments on two real-world datasets. Empirical results show that SML not only achieves significant speed-up, but also outperforms the full model retraining in recommendation accuracy, validating the effectiveness of our proposals. We release our codes at: https://github.com/zyang1580/SML.

References

  1. Alex Beutel, Paul Covington, Sagar Jain, Can Xu, Jia Li, Vince Gatto, and Ed H Chi. 2018. Latent cross: Making use of context in recurrent recommender systems. In WSDM. 46--54.Google ScholarGoogle Scholar
  2. Homanga Bharadhwaj. 2019. Meta-Learning for User Cold-Start Recommendation. In IJCNN. 1--8.Google ScholarGoogle Scholar
  3. Shiyu Chang, Yang Zhang, Jiliang Tang, Dawei Yin, Yi Chang, Mark A Hasegawa-Johnson, and Thomas S Huang. 2017. Streaming recommender systems. In WWW. 525--534.Google ScholarGoogle Scholar
  4. Yihong Chen, Bei Chen, Xiangnan He, Chen Gao, Yong Li, Jian-Guang Lou, and Yue Wang. 2019. (łambda)Opt: Learn to Regularize Recommender Models in Finer Levels. In SIGKDD. 978--986.Google ScholarGoogle Scholar
  5. Robin Devooght, Nicolas Kourtellis, and Amin Mantrach. 2015. Dynamic matrix factorization with priors on unknown values. In SIGKDD. 189--198.Google ScholarGoogle Scholar
  6. Ernesto Diaz-Aviles, Lucas Drumond, Lars Schmidt-Thieme, and Wolfgang Nejdl. 2012. Real-time top-n recommendation in social streams. In RecSys. 59--66.Google ScholarGoogle Scholar
  7. Xiaoyu Du, Xiangnan He, Fajie Yuan, Jinhui Tang, Zhiguang Qin, and Tat-Seng Chua. 2019 a. Modeling Embedding Dimension Correlations via Convolutional Neural Collaborative Filtering. TOIS, Vol. 37, 4 (2019), 47:1--47:22.Google ScholarGoogle ScholarDigital LibraryDigital Library
  8. Zhengxiao Du, Xiaowei Wang, Hongxia Yang, Jingren Zhou, and Jie Tang. 2019 b. Sequential Scenario-Specific Meta Learner for Online Recommendation. In SIGKDD. 2895--2904.Google ScholarGoogle Scholar
  9. Chelsea Finn, Pieter Abbeel, and Sergey Levine. 2017. Model-agnostic meta-learning for fast adaptation of deep networks. In ICML, Vol. 70. 1126--1135.Google ScholarGoogle Scholar
  10. Luca Franceschi, Paolo Frasconi, Saverio Salzo, Riccardo Grazzi, and Massimiliano Pontil. 2018. Bilevel Programming for Hyperparameter Optimization and Meta-Learning. In ICML, Vol. 80. 1563--1572.Google ScholarGoogle Scholar
  11. Jon Atle Gulla, Lemei Zhang, Peng Liu, Özlem Özgöbek, and Xiaomeng Su. 2017. The Adressa dataset for news recommendation. In WI. 1042--1048.Google ScholarGoogle Scholar
  12. Ruining He, Wang-Cheng Kang, and Julian McAuley. 2017a. Translation-based recommendation. In RecSys. 161--169.Google ScholarGoogle Scholar
  13. Xiangnan He and Tat-Seng Chua. 2017. Neural factorization machines for sparse predictive analytics. In SIGIR. 355--364.Google ScholarGoogle Scholar
  14. Xiangnan He, Kuan Deng, Xiang Wang, Yan Li, Yongdong Zhang, and Meng Wang. 2020. LightGCN: Simplifying and Powering Graph Convolution Network for Recommendation. In SIGIR.Google ScholarGoogle Scholar
  15. Xiangnan He, Lizi Liao, Hanwang Zhang, Liqiang Nie, Xia Hu, and Tat-Seng Chua. 2017b. Neural collaborative filtering. In WWW. 173--182.Google ScholarGoogle Scholar
  16. Xiangnan He, Hanwang Zhang, Min-Yen Kan, and Tat-Seng Chua. 2016. Fast Matrix Factorization for Online Recommendation with Implicit Feedback. In SIGIR. 549--558.Google ScholarGoogle Scholar
  17. Dan Hendrycks and Kevin Gimpel. 2016. Bridging Nonlinearities and Stochastic Regularizers with Gaussian Error Linear Units. CoRR, Vol. abs/1606.08415 (2016).Google ScholarGoogle Scholar
  18. Balá zs Hidasi, Alexandros Karatzoglou, Linas Baltrunas, and Domonkos Tikk. 2016. Session-based Recommendations with Recurrent Neural Networks. In ICLR.Google ScholarGoogle Scholar
  19. Kurt Hornik. 1991. Approximation capabilities of multilayer feedforward networks. Neural networks, Vol. 4, 2 (1991), 251--257.Google ScholarGoogle Scholar
  20. Haoji Hu, Xiangnan He, Jinyang Gao, and Zhi-Li Zhang. 2020. Modeling Personalized Item Frequency Information for Next-basket Recommendation. In SIGIR.Google ScholarGoogle Scholar
  21. Muhammad Abdullah Jamal and Guo-Jun Qi. 2019. Task Agnostic Meta-Learning for Few-Shot Learning. In CVPR. 11719--11727.Google ScholarGoogle Scholar
  22. Diederik P. Kingma and Jimmy Ba. 2015. Adam: A Method for Stochastic Optimization. In ICLR.Google ScholarGoogle Scholar
  23. James Kirkpatrick, Razvan Pascanu, Neil C Rabinowitz, Joel Veness, Guillaume Desjardins, Andrei A Rusu, Kieran Milan, John Quan, Tiago Ramalho, Agnieszka Grabskabarwinska, et almbox. 2017. Overcoming catastrophic forgetting in neural networks. PNAS, Vol. 114, 13 (2017), 3521--3526.Google ScholarGoogle ScholarCross RefCross Ref
  24. Hoyeop Lee, Jinbae Im, Seongwon Jang, Hyunsouk Cho, and Sehee Chung. 2019. MeLU: Meta-Learned User Preference Estimator for Cold-Start Recommendation. In SIGKDD. 1073--1082.Google ScholarGoogle Scholar
  25. Wenqiang Lei, Xiangnan He, Yisong Miao, Qingyun Wu, Richang Hong, Min-Yen Kan, and Tat-Seng Chua. 2020. Estimation-Action-Reflection: Towards Deep Interaction Between Conversational and Recommender Systems. In WSDM. 304--312.Google ScholarGoogle Scholar
  26. Jianxun Lian, Xiaohuan Zhou, Fuzheng Zhang, Zhongxia Chen, Xing Xie, and Guangzhong Sun. 2018. xdeepfm: Combining explicit and implicit feature interactions for recommender systems. In SIGKDD. 1754--1763.Google ScholarGoogle Scholar
  27. David Lopez-Paz and Marc'Aurelio Ranzato. 2017. Gradient Episodic Memory for Continual Learning. In NeurlPS 2017. 6467--6476.Google ScholarGoogle Scholar
  28. Feiyang Pan, Shuokai Li, Xiang Ao, Pingzhong Tang, and Qing He. 2019. Warm Up Cold-start Advertisements: Improving CTR Predictions via Learning to Learn ID Embeddings. In SIGIR. 695--704.Google ScholarGoogle Scholar
  29. Massimo Quadrana, Paolo Cremonesi, and Dietmar Jannach. 2018. Sequence-aware recommender systems. ACM Computing Surveys (CSUR), Vol. 51, 4 (2018), 66:1--66:36.Google ScholarGoogle ScholarDigital LibraryDigital Library
  30. Sachin Ravi and Hugo Larochelle. 2017. Optimization as a Model for Few-Shot Learning. In ICLR.Google ScholarGoogle Scholar
  31. Yi Ren, Cuirong Chi, and Zhang Jintao. 2019. A Survey of Personalized Recommendation Algorithm Selection Based on Meta-learning. In CSIA. 1383--1388.Google ScholarGoogle Scholar
  32. Steffen Rendle. 2010. Factorization Machines. In ICDM. 995--1000.Google ScholarGoogle Scholar
  33. Steffen Rendle, Christoph Freudenthaler, Zeno Gantner, and Lars Schmidt-Thieme. 2009. BPR: Bayesian personalized ranking from implicit feedback. In UAI. 452--461.Google ScholarGoogle ScholarDigital LibraryDigital Library
  34. Steffen Rendle, Christoph Freudenthaler, and Lars Schmidt-Thieme. 2010. Factorizing personalized markov chains for next-basket recommendation. In WWW. 811--820.Google ScholarGoogle Scholar
  35. Steffen Rendle and Lars Schmidt-Thieme. 2008. Online-updating regularized kernel matrix factorization models for large-scale recommender systems. In RecSys. 251--258.Google ScholarGoogle Scholar
  36. David Sculley, Gary Holt, Daniel Golovin, Eugene Davydov, Todd Phillips, Dietmar Ebner, Vinay Chaudhary, Michael Young, Jean-Francois Crespo, and Dan Dennison. 2015. Hidden technical debt in machine learning systems. In NeurlPS. 2503--2511.Google ScholarGoogle Scholar
  37. Karthik Subbian, Charu Aggarwal, and Kshiteesh Hegde. 2016. Recommendations for streaming data. In CIKM. 2185--2190.Google ScholarGoogle Scholar
  38. Jiaxi Tang and Ke Wang. 2018. Personalized top-n sequential recommendation via convolutional sequence embedding. In WSDM. 565--573.Google ScholarGoogle Scholar
  39. Manasi Vartak, Arvind Thiagarajan, Conrado Miranda, Jeshua Bratman, and Hugo Larochelle. 2017. A meta-learning perspective on cold-start recommendations for items. In NeurlPS. 6904--6914.Google ScholarGoogle Scholar
  40. Jeffrey S Vitter. 1985. Random sampling with a reservoir. TOMS, Vol. 11, 1, 37--57.Google ScholarGoogle ScholarDigital LibraryDigital Library
  41. Qinyong Wang, Hongzhi Yin, Zhiting Hu, Defu Lian, Hao Wang, and Zi Huang. 2018a. Neural memory streaming recommender networks with adversarial training. In SIGKDD. 2467--2475.Google ScholarGoogle Scholar
  42. Weiqing Wang, Hongzhi Yin, Zi Huang, Qinyong Wang, Xingzhong Du, and Quoc Viet Hung Nguyen. 2018b. Streaming ranking based recommender systems. In SIGIR. 525--534.Google ScholarGoogle Scholar
  43. Xiang Wang, Xiangnan He, Yixin Cao, Meng Liu, and Tat-Seng Chua. 2019 a. KGAT: Knowledge Graph Attention Network for Recommendation. In SIGKDD. 950--958.Google ScholarGoogle Scholar
  44. Xiang Wang, Xiangnan He, Meng Wang, Fuli Feng, and Tat-Seng Chua. 2019 b. Neural Graph Collaborative Filtering. In SIGIR. 165--174.Google ScholarGoogle Scholar
  45. Bin Wu, Xiangnan He, Zhongchuan Sun, Liang Chen, and Yangdong Ye. 2019. ATM: An Attentive Translation Model for Next-Item Recommendation. IEEE Transactions on Industrial Informatics, Vol. 16, 3 (2019), 1448--1459.Google ScholarGoogle ScholarCross RefCross Ref
  46. Fajie Yuan, Xiangnan He, Haochuan Jiang, Guibing Guo, Jian Xiong, Zhezhao Xu, and Yilin Xiong. 2020. Future Data Helps Training: Modeling Future Contexts for Session-based Recommendation. In WWW. 303--313.Google ScholarGoogle Scholar
  47. Fajie Yuan, Alexandros Karatzoglou, Ioannis Arapakis, Joemon M Jose, and Xiangnan He. 2019. A Simple Convolutional Generative Network for Next Item Recommendation. In WSDM. 582--590.Google ScholarGoogle Scholar

Index Terms

  1. How to Retrain Recommender System?: A Sequential Meta-Learning Method

    Recommendations

    Comments

    Login options

    Check if you have access through your login credentials or your institution to get full access on this article.

    Sign in
    • Published in

      cover image ACM Conferences
      SIGIR '20: Proceedings of the 43rd International ACM SIGIR Conference on Research and Development in Information Retrieval
      July 2020
      2548 pages
      ISBN:9781450380164
      DOI:10.1145/3397271

      Copyright © 2020 ACM

      Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than the author(s) must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected].

      Publisher

      Association for Computing Machinery

      New York, NY, United States

      Publication History

      • Published: 25 July 2020

      Permissions

      Request permissions about this article.

      Request Permissions

      Check for updates

      Qualifiers

      • research-article

      Acceptance Rates

      Overall Acceptance Rate792of3,983submissions,20%

    PDF Format

    View or Download as a PDF file.

    PDF

    eReader

    View online with eReader.

    eReader