Reinforcement Learning Agents for Interacting with Humans
Skip to main content
eScholarship
Open Access Publications from the University of California

Reinforcement Learning Agents for Interacting with Humans

Abstract

We tackle the problem of an agent interacting with humans in a general-sum environment, i.e., a non-zero sum, non-fully cooperative setting, where the agent's goal is to increase its own utility. We show that when data is limited, building an accurate human model is very challenging, and that a reinforcement learning agent, which is based on this data, does not perform well in practice. Therefore, we propose that the agent should try maximizing a linear combination of the human's utility and its own utility rather than simply trying to maximize only its own utility.

Main Content
For improved accessibility of PDF content, download the file to your device.
Current View