Open this publication in new window or tab >>2018 (English)In: Proceeding IUI '18 23rd International Conference on Intelligent User Interfaces, ACM Digital Library, 2018, p. 19-29Conference paper, Published paper (Refereed)
Abstract [en]
This work presents an extension of Thompson Sampling bandit policy for orchestrating the collection of base recommendation algorithms for e-commerce. We focus on the problem of item-to-item recommendations, for which multiple behavioral and attribute-based predictors are provided to an ensemble learner. We show how to adapt Thompson Sampling to realistic situations when neither action availability nor reward stationarity is guaranteed. Furthermore, we investigate the effects of priming the sampler with pre-set parameters of reward probability distributions by utilizing the product catalog and/or event history, when such information is available. We report our experimental results based on the analysis of three real-world e-commerce datasets.
Place, publisher, year, edition, pages
ACM Digital Library, 2018
Keywords
recommender system, e-commerce, thompson sampling
National Category
Engineering and Technology
Identifiers
urn:nbn:se:mau:diva-12346 (URN)10.1145/3172944.3172967 (DOI)000458192600005 ()2-s2.0-85045442051 (Scopus ID)27340 (Local ID)27340 (Archive number)27340 (OAI)
Conference
23rd Intetnational Conference on Intelligent User Interface (IUI23), Tokyo, Japan (March 7-11, 2018)
2020-02-292020-02-292024-06-17Bibliographically approved