| | |
| | |
Stat |
Members: 3645 Articles: 2'503'724 Articles rated: 2609
24 April 2024 |
|
| | | |
|
Article overview
| |
|
Fitting a Linear Control Policy to Demonstrations with a Kalman Constraint | Malayandi Palan
; Shane Barratt
; Alex McCauley
; Dorsa Sadigh
; Vikas Sindhwani
; Stephen Boyd
; | Date: |
21 Jan 2020 | Abstract: | We consider the problem of learning a linear control policy for a linear
dynamical system, from demonstrations of an expert regulating the system. The
standard approach to this problem is policy fitting, which fits a linear policy
by minimizing a loss function between the demonstrations and the policy’s
outputs plus a regularization function that encodes prior knowledge. Despite
its simplicity, this method fails to learn policies with low or even finite
cost when there are few demonstrations. We propose to add an additional
constraint to policy fitting, that the policy is the solution to some LQR
problem, i.e., optimal in the stochastic control sense for some choice of
quadratic cost. We refer to this constraint as a Kalman constraint. Policy
fitting with a Kalman constraint requires solving an optimization problem with
convex cost and bilinear constraints. We propose a heuristic method, based on
the alternating direction method of multipliers (ADMM), to approximately solve
this problem. Numerical experiments demonstrate that adding the Kalman
constraint allows us to learn good, i.e., low cost, policies even when very few
data are available. | Source: | arXiv, 2001.7572 | Services: | Forum | Review | PDF | Favorites |
|
|
No review found.
Did you like this article?
Note: answers to reviews or questions about the article must be posted in the forum section.
Authors are not allowed to review their own article. They can use the forum section.
browser Mozilla/5.0 AppleWebKit/537.36 (KHTML, like Gecko; compatible; ClaudeBot/1.0; +claudebot@anthropic.com)
|
| |
|
|
|
| News, job offers and information for researchers and scientists:
| |