Skip to main content
eScholarship
Open Access Publications from the University of California

Deep Reinforcement Learning in Buildings: Implicit Assumptions and their Impact

  • Author(s): Prakash, AK;
  • Touzani, S;
  • Kiran, M;
  • Agarwal, S;
  • Pritoni, M;
  • Granderson, J
  • et al.
Abstract

As deep reinforcement learning (DRL) continues to gain interest in the smart building research community, there is a transition from simulation-based evaluations to deploying DRL control strategies in actual buildings. While the efficacy of a solution could depend on a particular implementation, there are common obstacles that developers have to overcome to deliver an effective controller. Additionally, a deployment in a physical building can invalidate some of the assumptions made during the controller development. Assumptions on the sensor placement or on the equipment behavior can quickly come undone. This paper presents some of the significant assumptions made during the development of DRL based controllers that could affect their operations in a physical building. Furthermore, a preliminary evaluation revealed that controllers developed with some of these assumptions can incur twice the expected costs when they are deployed in a building.

Main Content
For improved accessibility of PDF content, download the file to your device.
Current View