Rethinking Reinforcement Learning for Recommendation: A Prompt Perspective
Published in Proceedings of the 45th International ACM SIGIR Conference on Research and Development in Information Retrieval, 2022
Modern recommender systems aim to improve user experience. As reinforcement learning (RL) naturally fits this objective – maximizing an user's reward per session – it has become an emerging topic in recommender systems. Developing RL-based recommendation methods, however, is not trivial due to the \emph{offline training challenge}. Specifically, the keystone of traditional RL is to train an agent with large amounts of online exploration making lots of errors' in the process. In the recommendation setting, though, we cannot afford the price of making
errors' online. As a result, the agent needs to be trained through offline historical implicit feedback, collected under different recommendation policies; traditional RL algorithms may lead to sub-optimal policies under these offline training settings. Here we propose a new learning paradigm – namely Prompt-Based Reinforcement Learning (PRL) – for the offline training of RL-based recommendation agents. While traditional RL algorithms attempt to map state-action input pairs to their expected rewards (e.g., Q-values), PRL directly infers actions (i.e., recommended items) from state-reward inputs. In short, the agents are trained to predict a recommended item given the prior interactions and an observed reward value – with simple supervised learning. At deployment time, this historical (training) data acts as a knowledge base, while the state-reward pairs are used as a prompt. The agents are thus used to answer the question: \emph{ Which item should be recommended given the prior interactions \& the prompted reward value}? We implement PRL with four notable recommendation models and conduct experiments on two real-world e-commerce datasets. Experimental results demonstrate the superior performance of our proposed methods.
@inproceedings{xin-etal-2022-rethinking,
author = {
Xin Xin and
Tiago Pimentel and
Alexandros Karatzoglou and
Pengjie Ren and
Konstantina Christakopoulou and
Zhaochun Ren
},
booktitle = {Proceedings of the 45th International ACM SIGIR Conference on Research and Development in Information Retrieval},
title = {Rethinking Reinforcement Learning for Recommendation: A Prompt Perspective},
year = {2022},
doi = {10.1145/3477495.3531714},
url = {https://dl.acm.org/doi/10.1145/3477495.3531714},
pages = {1347--1357},
}