In various settings time consistency in dynamic programming has been addressed by many authors going all the way back to original developments by Richard Bellman. The basic idea of the involved dynamic principle is that a policy designed at the first stage, before observing realizations of the random data, should not be changed at the later stages of the decision process. This is a rather vague principle since this leaves a choice of optimality criteria at every stage of the process conditional on an observed realization of the random data. In this paper we discuss this from the point of view of modern theory of risk averse stochastic programming. In particular we discuss time consistent decision making by addressing risk measures which are recursive, nested, dynamically or time consistent. It turns out that the paradigm of time consistency is in conflict with various desirable, classical properties of general risk measures.
School of Industrial and Systems Engineering, Georgia Institute of Technology, Atlanta, GA 30332-0205 Faculty of Mathematics, Chemnitz University of Technology, D-09107 Chemnitz, Germany