skip to main content
10.1145/3383313.3418485acmconferencesArticle/Chapter ViewAbstractPublication PagesrecsysConference Proceedingsconference-collections
extended-abstract

A Joint Dynamic Ranking System with DNN and Vector-based Clustering Bandit

Published: 22 September 2020 Publication History

Abstract

The ad-ranking module is the core of the advertising recommender system. Existing ad-ranking modules are mainly based on the deep neural network click-through rate prediction model. Recently an innovative ad-ranking paradigm called DNN-MAB has been introduced to address DNN-only paradigms’ weakness in perceiving highly dynamic user intent over time. We introduce the DNN-MAB paradigm into our ad-ranking system to alleviate the Matthew effect that harms the user experience. Due to data sparsity, however, the actual performance of DNN-MAB is lower than expected. In this paper, we propose an innovative ad-ranking paradigm called DNN-VMAB to solve these problems. Based on vectorization and clustering, it utilizes latent collaborative information in user behavior data to find a set of ads with higher relativity and diversity. As an integration of the essences of classical collaborative filtering, deep click-through rate prediction model, and contextual multi-armed bandit, it can improve platform revenue and user experience. Both offline and online experiments show the advantage of our new algorithm over DNN-MAB and some other existing algorithms.

References

[1]
Peter Auer. 2002. Using confidence bounds for exploitation-exploration trade-offs. Journal of Machine Learning Research 3, Nov (2002), 397–422.
[2]
Guy Bresler, George H Chen, and Devavrat Shah. 2014. A latent source model for online collaborative filtering. In Advances in Neural Information Processing Systems. 3347–3355.
[3]
Yufei Feng, Fuyu Lv, Weichen Shen, Menghan Wang, Fei Sun, Yu Zhu, and Keping Yang. 2019. Deep Session Interest Network for Click-Through Rate Prediction. In Proceedings of the Twenty-Eighth International Joint Conference on Artificial Intelligence. 2301–2307.
[4]
Claudio Gentile, Shuai Li, Purushottam Kar, Alexandros Karatzoglou, Giovanni Zappella, and Evans Etrue. 2017. On context-dependent clustering of bandits. In Proceedings of the 34th International Conference on Machine Learning-Volume 70. JMLR. org, 1253–1262.
[5]
Claudio Gentile, Shuai Li, and Giovanni Zappella. 2014. Online clustering of bandits. In International Conference on Machine Learning. 757–765.
[6]
Branislav Kveton, Csaba Szepesvari, Zheng Wen, and Azin Ashkan. 2015. Cascading bandits: Learning to rank in the cascade model. In International Conference on Machine Learning. 767–776.
[7]
Chang Li, Branislav Kveton, Tor Lattimore, Ilya Markov, Maarten de Rijke, Csaba Szepesvári, and Masrour Zoghi. 2019. BubbleRank: Safe Online Learning to Re-Rank via Implicit Click Feedback. In Proceedings of the 35th Conference on Uncertainty in Artificial Intelligence. 47.
[8]
Shuai Li, Claudio Gentile, and Alexandros Karatzoglou. 2016. Graph clustering bandits for recommendation. arXiv preprint arXiv:1605.00596(2016).
[9]
Shuai Li, Alexandros Karatzoglou, and Claudio Gentile. 2016. Collaborative filtering bandits. In Proceedings of the 39th International ACM SIGIR conference on Research and Development in Information Retrieval. ACM, 539–548.
[10]
Yudan Liu, Kaikai Ge, Xu Zhang, and Leyu Lin. 2019. Real-time Attention Based Look-alike Model for Recommender System. In Proceedings of the 25th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining. 2765–2773.
[11]
Tomas Mikolov, Ilya Sutskever, Kai Chen, Greg S Corrado, and Jeff Dean. 2013. Distributed representations of words and phrases and their compositionality. In Advances in neural information processing systems. 3111–3119.
[12]
Trong T Nguyen and Hady W Lauw. 2014. Dynamic clustering of contextual multi-armed bandits. In Proceedings of the 23rd ACM International Conference on Conference on Information and Knowledge Management. ACM, 1959–1962.
[13]
Javier Sanz-Cruzado, Pablo Castells, and Esther López. 2019. A simple multi-armed nearest-neighbor bandit for interactive recommendation. In Proceedings of the 13th ACM Conference on Recommender Systems. ACM, 358–362.
[14]
Aleksandrs Slivkins, Filip Radlinski, and Sreenivas Gollapudi. 2013. Ranked bandits in metric spaces: learning diverse rankings over large document collections. Journal of Machine Learning Research 14, Feb (2013), 399–436.
[15]
Stephan Ten Hagen, Maarten Van Someren, Vera Hollink, 2003. Exploration/exploitation in adaptive recommender systems. proceedings of Eunite 2003(2003).
[16]
Ruoxi Wang, Bin Fu, Gang Fu, and Mingliang Wang. 2017. Deep & cross network for ad click predictions. In Proceedings of the ADKDD’17. ACM, 12.
[17]
Christopher John Cornish Hellaby Watkins. 1989. Learning from Delayed Rewards. Ph.D. Dissertation. King’s College, Cambridge, UK. http://www.cs.rhul.ac.uk/~chrisw/new_thesis.pdf
[18]
Yan Yan, Zitao Liu, Meng Zhao, Wentao Guo, Weipeng P Yan, and Yongjun Bao. 2018. A Practical Deep Online Ranking System in E-commerce Recommendation. In Joint European Conference on Machine Learning and Knowledge Discovery in Databases. Springer, 186–201.

Index Terms

  1. A Joint Dynamic Ranking System with DNN and Vector-based Clustering Bandit
          Index terms have been assigned to the content through auto-classification.

          Recommendations

          Comments

          Information & Contributors

          Information

          Published In

          cover image ACM Conferences
          RecSys '20: Proceedings of the 14th ACM Conference on Recommender Systems
          September 2020
          796 pages
          ISBN:9781450375832
          DOI:10.1145/3383313
          Permission to make digital or hard copies of part or all of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for third-party components of this work must be honored. For all other uses, contact the Owner/Author.

          Sponsors

          Publisher

          Association for Computing Machinery

          New York, NY, United States

          Publication History

          Published: 22 September 2020

          Check for updates

          Author Tags

          1. Learning to rank
          2. Multi-arm bandits

          Qualifiers

          • Extended-abstract
          • Research
          • Refereed limited

          Conference

          RecSys '20: Fourteenth ACM Conference on Recommender Systems
          September 22 - 26, 2020
          Virtual Event, Brazil

          Acceptance Rates

          Overall Acceptance Rate 254 of 1,295 submissions, 20%

          Upcoming Conference

          RecSys '24
          18th ACM Conference on Recommender Systems
          October 14 - 18, 2024
          Bari , Italy

          Contributors

          Other Metrics

          Bibliometrics & Citations

          Bibliometrics

          Article Metrics

          • 0
            Total Citations
          • 273
            Total Downloads
          • Downloads (Last 12 months)9
          • Downloads (Last 6 weeks)0
          Reflects downloads up to 14 Sep 2024

          Other Metrics

          Citations

          View Options

          Get Access

          Login options

          View options

          PDF

          View or Download as a PDF file.

          PDF

          eReader

          View online with eReader.

          eReader

          HTML Format

          View this article in HTML Format.

          HTML Format

          Media

          Figures

          Other

          Tables

          Share

          Share

          Share this Publication link

          Share on social media