Linear Bayes policy for learning in contextual-bandits
作者:
Highlights:
• A new selection rule for Contextual Bandits and single-step Reinforcement Learning.
• The use of empirical approximations to the Bayes’rule is an effective approach.
• This technique arrived second at new Challenges for Exploration & Exploitation 3.
摘要
•A new selection rule for Contextual Bandits and single-step Reinforcement Learning.•The use of empirical approximations to the Bayes’rule is an effective approach.•This technique arrived second at new Challenges for Exploration & Exploitation 3.
论文关键词:Contextual bandits,Online advertising,Recommender systems,One-to-one Marketing,Empirical Bayes
论文评审过程:Available online 18 July 2013.
论文官网地址:https://doi.org/10.1016/j.eswa.2013.07.041