Dynamic programming approach to principal-agent problems
Jak\v{s}a Cvitani\'c,
Dylan Possama\"i and
Nizar Touzi
Papers from arXiv.org
Abstract:
We consider a general formulation of the Principal-Agent problem with a lump-sum payment on a finite horizon, providing a systematic method for solving such problems. Our approach is the following: we first find the contract that is optimal among those for which the agent's value process allows a dynamic programming representation, for which the agent's optimal effort is straightforward to find. We then show that the optimization over the restricted family of contracts represents no loss of generality. As a consequence, we have reduced this non-zero sum stochastic differential game to a stochastic control problem which may be addressed by the standard tools of control theory. Our proofs rely on the backward stochastic differential equations approach to non-Markovian stochastic control, and more specifically, on the recent extensions to the second order case.
Date: 2015-10, Revised 2017-01
References: View references in EconPapers View complete reference list from CitEc
Citations: View citations in EconPapers (6)
Downloads: (external link)
http://arxiv.org/pdf/1510.07111 Latest version (application/pdf)
Related works:
This item may be available elsewhere in EconPapers: Search for items with the same title.
Export reference: BibTeX
RIS (EndNote, ProCite, RefMan)
HTML/Text
Persistent link: https://EconPapers.repec.org/RePEc:arx:papers:1510.07111
Access Statistics for this paper
More papers in Papers from arXiv.org
Bibliographic data for series maintained by arXiv administrators ().