Dynamic programming approach to principal–agent problems
- Creators
-
Cvitanić, Jakša
- Possamaï, Dylan
- Touzi, Nizar
Abstract
We consider a general formulation of the principal–agent problem with a lump-sum payment on a finite horizon, providing a systematic method for solving such problems. Our approach is the following. We first find the contract that is optimal among those for which the agent's value process allows a dynamic programming representation, in which case the agent's optimal effort is straightforward to find. We then show that the optimization over this restricted family of contracts represents no loss of generality. As a consequence, we have reduced a non-zero-sum stochastic differential game to a stochastic control problem which may be addressed by standard tools of control theory. Our proofs rely on the backward stochastic differential equations approach to non-Markovian stochastic control, and more specifically on the recent extensions to the second order case.
Additional Information
© 2017 Springer-Verlag GmbH Germany. Received: 19 April 2016; Accepted: 27 March 2017; First Online: 27 October 2017.Attached Files
Submitted - 1510.07111.pdf
Files
Name | Size | Download all |
---|---|---|
md5:3300a0ca3aa0e1e8571051e0779e827e
|
352.9 kB | Preview Download |
Additional details
- Eprint ID
- 82728
- DOI
- 10.1007/s00780-017-0344-4
- Resolver ID
- CaltechAUTHORS:20171027-093026885
- Created
-
2017-10-27Created from EPrint's datestamp field
- Updated
-
2021-11-15Created from EPrint's last_modified field