Institute of Information Theory and Automation

Publication details

Risk-Sensitive and Mean Variance Optimality in Markov Decision Processes

Journal Article

Sladký Karel


serial: Acta Oeconomica Pragensia vol.7, 3 (2013), p. 146-161

project(s): 171396, AVČR a CONACyT, GAP402/10/0956, GA ČR, GAP402/11/0150, GA ČR

keywords: Discrete-time Markov decision chains, exponential utility functions, certainty equivalent, mean-variance optimality, connections between risk-sensitive and risk-neutral models

preview: Download

abstract (eng):

In this paper we consider unichain Markov decision processes with finite state space and compact actions spaces where the stream of rewards generated by the Markov processes is evaluated by an exponential utility function with a given risk sensitivity coefficient (so-called risk-sensitive models). If the risk sensitivity coefficient equals zero (risk-neutral case) we arrive at a standard Markov decision process. Then we can easily obtain necessary and sufficient mean reward optimality conditions and the variability can be evaluated by the mean variance of total expected rewards. For the risk-sensitive case we establish necessary and sufficient optimality conditions for maximal (or minimal) growth rate of expectation of the exponential utility function,¨along with mean value of the corresponding certainty equivalent, that take into account not only the expected values of the total reward but also its higher moments.

RIV: BB

Responsible for information: admin
Last modification: 21.12.2012
Institute of Information Theory and Automation