Constraint Inference in Control and Reinforcement Learning
Skip to main content
eScholarship
Open Access Publications from the University of California

UC Berkeley

UC Berkeley Electronic Theses and Dissertations bannerUC Berkeley

Constraint Inference in Control and Reinforcement Learning

Abstract

Inferring unknown constraints is a challenging and crucial problem in many robotics applications. When only expert demonstrations are available, it becomes essential to inferthe unknown domain constraints to deploy additional agents effectively. In this work, we propose approaches to infer constraints by observing experts act in an environment in different scenarios.

First, we develop an approach to infer affine constraints in control tasks by observing expertdemonstrations. We formulate the constraint inference problem as an inverse optimization problem, and we propose an alternating optimization scheme that infers the unknown constraints by minimizing a KKT residual objective. We demonstrate the effectiveness of our method in a number of simulations, and show that our method can infer less conservative constraints than another baseline method, while maintaining comparable safety guarantees.

Second, we move to a Reinforcement Learning framework and we consider the problem ofinferring constraints from demonstrations using a Bayesian perspective. We propose Bayesian Inverse Constraint Reinforcement Learning (BICRL), a novel approach that infers a posterior probability distribution over constraints from demonstrated trajectories. The main advantages of BICRL, compared to prior constraint inference algorithms, are (1) the freedom to infer constraints from partial trajectories and even from disjoint state-action pairs, (2) the ability to infer constraints from suboptimal demonstrations and in stochastic environments, and (3) the opportunity to use the posterior distribution over constraints in order to implement active learning and robust policy optimization techniques. We show that BICRL outperforms pre-existing constraint learning approaches, leading to more accurate constraint inference and consequently safer policies. We further propose Hierarchical BICRL that infers constraints locally in sub-spaces of the entire domain and then composes global constraint estimates leading to accurate and computationally efficient constraint estimation.

Our third contribution is also based on a Reinforcement Learning framework. In that, we propose a novel Bayesian method that infers constraints based on preferences overdemonstrations. The main advantages of our proposed approach are that it (1) infers constraints without calculating a new policy at each iteration, (2) uses a simple and more realistic ranking of groups of demonstrations, without requiring pairwise comparisons over all demonstrations, and (3) adapts to cases where there are varying levels of constraint violation. Our empirical results demonstrate that our proposed Bayesian approach infers constraints of varying severity, more accurately than state-of-the-art constraint inference methods.

Main Content
For improved accessibility of PDF content, download the file to your device.
Current View