Published July 2020
| Submitted
Book Section - Chapter
Open
Risk-Averse Planning Under Uncertainty
Chicago
Abstract
We consider the problem of designing policies for partially observable Markov decision processes (POMDPs) with dynamic coherent risk objectives. Synthesizing risk-averse optimal policies for POMDPs requires infinite memory and thus undecidable. To overcome this difficulty, we propose a method based on bounded policy iteration for designing stochastic but finite state (memory) controllers, which takes advantage of standard convex optimization methods. Given a memory budget and optimality criterion, the proposed method modifies the stochastic finite state controller leading to sub-optimal solutions with lower coherent risk.
Additional Information
© 2020 AACC.Attached Files
Submitted - 1909.12499.pdf
Files
1909.12499.pdf
Files
(1.8 MB)
Name | Size | Download all |
---|---|---|
md5:40b4b0150a782c9a3f66677dff933cc4
|
1.8 MB | Preview Download |
Additional details
- Eprint ID
- 100583
- Resolver ID
- CaltechAUTHORS:20200109-092433424
- Created
-
2020-01-09Created from EPrint's datestamp field
- Updated
-
2021-11-16Created from EPrint's last_modified field
- Caltech groups
- Division of Biology and Biological Engineering (BBE)