Dynamic consistency for stochastic optimal control problems

被引:20
作者
Carpentier, Pierre [1 ]
Chancelier, Jean-Philippe [2 ]
Cohen, Guy [2 ]
De Lara, Michel [2 ]
Girardeau, Pierre [1 ,2 ,3 ]
机构
[1] ENSTA ParisTech, F-75739 Paris 15, France
[2] Univ Paris Est, CERMICS, Ecole Ponts ParisTech, F-77455 Marne La Vallee 2, France
[3] EDF R&D, F-92141 Clamart, France
关键词
Stochastic optimal control; Dynamic consistency; Dynamic programming; Risk measures; RISK MEASURES; CHOICE; TASTES;
D O I
10.1007/s10479-011-1027-8
中图分类号
C93 [管理学]; O22 [运筹学];
学科分类号
070105 ; 12 ; 1201 ; 1202 ; 120202 ;
摘要
For a sequence of dynamic optimization problems, we aim at discussing a notion of consistency over time. This notion can be informally introduced as follows. At the very first time step t (0), the decision maker formulates an optimization problem that yields optimal decision rules for all the forthcoming time steps t (0),t (1),aEuro broken vertical bar,T; at the next time step t (1), he is able to formulate a new optimization problem starting at time t (1) that yields a new sequence of optimal decision rules. This process can be continued until the final time T is reached. A family of optimization problems formulated in this way is said to be dynamically consistent if the optimal strategies obtained when solving the original problem remain optimal for all subsequent problems. The notion of dynamic consistency, well-known in the field of economics, has been recently introduced in the context of risk measures, notably by Artzner et al. (Ann. Oper. Res. 152(1):5-22, 2007) and studied in the stochastic programming framework by Shapiro (Oper. Res. Lett. 37(3):143-147, 2009) and for Markov Decision Processes (MDP) by Ruszczynski (Math. Program. 125(2):235-261, 2010). We here link this notion with the concept of "state variable" in MDP, and show that a significant class of dynamic optimization problems are dynamically consistent, provided that an adequate state variable is chosen.
引用
收藏
页码:247 / 263
页数:17
相关论文
共 22 条
[1]  
[Anonymous], 1998, Variational Analysis
[2]  
[Anonymous], 2000, Dynamic programming and optimal control
[3]  
[Anonymous], 2013, Stochastic Programming
[4]   Coherent multiperiod risk adjusted values and Bellman's principle [J].
Artzner, Philippe ;
Delbaen, Freddy ;
Eber, Jean-Marc ;
Heath, David ;
Ku, Hyejin .
ANNALS OF OPERATIONS RESEARCH, 2007, 152 (1) :5-22
[5]  
Bellman R. E., 1957, Dynamic programming. Princeton landmarks in mathematics
[6]   Dynamic monetary risk measures for bounded discrete-time processes [J].
Cheridito, P ;
Delbaen, F ;
Krupper, M .
ELECTRONIC JOURNAL OF PROBABILITY, 2006, 11 :57-106
[7]   Conditional and dynamic convex risk measures [J].
Detlefsen, K ;
Scandolo, G .
FINANCE AND STOCHASTICS, 2005, 9 (04) :539-561
[8]   Richard Bellman on the birth of dynamic programming [J].
Dreyfus, S .
OPERATIONS RESEARCH, 2002, 50 (01) :48-51
[9]  
Ekeland I., 2006, ARXIVMATHOC0604264
[10]   CHANGING TASTES AND COHERENT DYNAMIC CHOICE [J].
HAMMOND, PJ .
REVIEW OF ECONOMIC STUDIES, 1976, 43 (01) :159-173