Abstract
Online recommendation task has been recognized as a Multi-Armed Bandit (MAB) problem. Despite the recent advances, there still needs to be more consensus on the best practices to evaluate such bandit solutions. Recently, we observed two complementary frameworks that allow us to evaluate bandit solutions more accurately: iRec and OBP. The first has a complete set of datasets, metrics, and MAB models implemented, allowing only offline evaluations of these solutions. However, the second is limited to a few bandit solutions with more current metrics and methodologies, such as counterfactuals. In this work, we propose and evaluate an integration between these two frameworks, demonstrating the potential and richness of analyzes that can be carried out from this combination.
This work was partially funded by CNPq, CAPES, FINEP and Fapemig.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Notes
- 1.
Available at https://github.com/YanAndrade61/iRec-OBP.
- 2.
Available at https://github.com/irec-org.
- 3.
Available at https://github.com/YanAndrade61/iRec-OBP.
References
Bobadilla, J., Ortega, F., Hernando, A., Gutiérrez, A.: Recommender systems survey. Knowl.-Based Syst. 46, 109–132 (2013)
Chapelle, O., Li, L.: An empirical evaluation of thompson sampling. In: Advances in Neural Information Processing Systems, pp. 2249–2257 (2011)
Liu, Y., Yen, J.N., Yuan, B., Shi, R., Yan, P., Lin, C.J.: Practical counterfactual policy learning for top-k recommendations. In: ACM SIGKDD, pp. 1141–1151 (2022)
Pan, W., Cui, S., Wen, H., Chen, K., Zhang, C., Wang, F.: Correcting the user feedback-loop bias for recommendation systems. arXiv preprint arXiv:2109.06037 (2021)
Saito, Y., Aihara, S., Matsutani, M., Narita, Y.: Open bandit dataset and pipeline: towards realistic and reproducible off-policy evaluation. arXiv preprint arXiv:2008.07146 (2020)
Sanz-Cruzado, J., Castells, P., López, E.: A simple multi-armed nearest-neighbor bandit for interactive recommendation. In: RecSys, pp. 358–362 (2019)
Shams, S., Anderson, D., Leith, D.: Cluster-based bandits: fast cold-start for recommender system new users (2021)
Silva, T., Silva, N., Werneck, H., Mito, C., Pereira, A.C., Rocha, L.: irec: an interactive recommendation framework. In: SIGIR, pp. 3165–3175 (2022)
Sutton, R.S., Barto, A.G.: Reinforcement Learning: An Introduction. MIT press, Cambridge (2018)
Wang, H., Wu, Q., Wang, H.: Factorization bandits for interactive recommendation. In: Thirty-First AAAI Conference on Artificial Intelligence (2017)
Wang, Q., et al.: Online interactive collaborative filtering using multi-armed bandit with dependent arms. IEEE Trans. Knowl. Data Eng. 31(8), 1569–1580 (2018)
Wu, Q., Iyer, N., Wang, H.: Learning contextual bandits in a non-stationary environment. In: SIGIR, pp. 495–504 (2018)
Yang, Y., Xia, X., Lo, D., Grundy, J.: A survey on deep learning for software engineering. ACM Comput. Surv. (CSUR) 54(10s), 1–73 (2022)
Zhao, X., Zhang, W., Wang, J.: Interactive collaborative filtering. In: Proceedings of the 22nd ACM International Conference on Information & Knowledge Management, pp. 1411–1420 (2013)
Zhou, S., et al.: Interactive recommender system via knowledge graph-enhanced reinforcement learning. In: SIGIR, pp. 179–188 (2020)
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2023 The Author(s), under exclusive license to Springer Nature Switzerland AG
About this paper
Cite this paper
Andrade, Y. et al. (2023). Integrating Counterfactual Evaluations into Traditional Interactive Recommendation Frameworks. In: Gervasi, O., et al. Computational Science and Its Applications – ICCSA 2023. ICCSA 2023. Lecture Notes in Computer Science, vol 13956 . Springer, Cham. https://doi.org/10.1007/978-3-031-36805-9_41
Download citation
DOI: https://doi.org/10.1007/978-3-031-36805-9_41
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-031-36804-2
Online ISBN: 978-3-031-36805-9
eBook Packages: Computer ScienceComputer Science (R0)