Dynamic programming approach to principal–agent problems

Jakša Cvitanić, Dylan Possamaï, Nizar Touzi

Research output: Contribution to journalArticlepeer-review

Abstract

We consider a general formulation of the principal–agent problem with a lump-sum payment on a finite horizon, providing a systematic method for solving such problems. Our approach is the following. We first find the contract that is optimal among those for which the agent’s value process allows a dynamic programming representation, in which case the agent’s optimal effort is straightforward to find. We then show that the optimization over this restricted family of contracts represents no loss of generality. As a consequence, we have reduced a non-zero-sum stochastic differential game to a stochastic control problem which may be addressed by standard tools of control theory. Our proofs rely on the backward stochastic differential equations approach to non-Markovian stochastic control, and more specifically on the recent extensions to the second order case.

Original languageEnglish (US)
Pages (from-to)1-37
Number of pages37
JournalFinance and Stochastics
Volume22
Issue number1
DOIs
StatePublished - Jan 1 2018

Keywords

  • Contract theory
  • Hamilton–Jacobi–Bellman equations
  • Principal–agent problem
  • Second order backward SDEs
  • Stochastic control of non-Markov systems

ASJC Scopus subject areas

  • Statistics and Probability
  • Finance
  • Statistics, Probability and Uncertainty

Fingerprint

Dive into the research topics of 'Dynamic programming approach to principal–agent problems'. Together they form a unique fingerprint.

Cite this