Approximate dynamic programming for optimal stationary control with control-dependent noise

Research output: Contribution to journalArticlepeer-review

Abstract

This brief studies the stochastic optimal control problem via reinforcement learning and approximate/adaptive dynamic programming (ADP). A policy iteration algorithm is derived in the presence of both additive and multiplicative noise using It calculus. The expectation of the approximated cost matrix is guaranteed to converge to the solution of some algebraic Riccati equation that gives rise to the optimal cost value. Moreover, the covariance of the approximated cost matrix can be reduced by increasing the length of time interval between two consecutive iterations. Finally, a numerical example is given to illustrate the efficiency of the proposed ADP methodology.

Original languageEnglish (US)
Article number6026952
Pages (from-to)2392-2398
Number of pages7
JournalIEEE Transactions on Neural Networks
Volume22
Issue number12 PART 2
DOIs
StatePublished - Dec 2011

Keywords

  • Approximate dynamic programming
  • control-dependent noise
  • optimal stationary control
  • stochastic systems

ASJC Scopus subject areas

  • Software
  • Computer Science Applications
  • Computer Networks and Communications
  • Artificial Intelligence

Fingerprint

Dive into the research topics of 'Approximate dynamic programming for optimal stationary control with control-dependent noise'. Together they form a unique fingerprint.

Cite this