Risk-Averse Stochastic Shortest Path Planning
Abstract
We consider the stochastic shortest path planning problem in MDPs, i.e., the problem of designing policies that ensure reaching a goal state from a given initial state with minimum accrued cost. In order to account for rare but important realizations of the system, we consider a nested dynamic coherent risk total cost functional rather than the conventional risk-neutral total expected cost. Under some assumptions, we show that optimal, stationary, Markovian policies exist and can be found via a special Bellman's equation. We propose a computational technique based on difference convex programs (DCPs) to find the associated value functions and therefore the risk-averse policies. A rover navigation MDP is used to illustrate the proposed methodology with conditional-value-at-risk (CVaR) and entropic-value-at-risk (EVaR) coherent risk measures.
Additional Information
© 2021 IEEE.Attached Files
Submitted - 2103.14727.pdf
Files
Name | Size | Download all |
---|---|---|
md5:b69e84986e92fa5781eb9d59f38bfabd
|
400.7 kB | Preview Download |
Additional details
- Eprint ID
- 109065
- DOI
- 10.1109/CDC45484.2021.9683527
- Resolver ID
- CaltechAUTHORS:20210511-082139184
- Created
-
2021-05-11Created from EPrint's datestamp field
- Updated
-
2022-02-16Created from EPrint's last_modified field