Markov Decision Models with Weighted Discounted Criteria
Author(s) -
Eugene A. Feinberg,
Adam Shwartz
Publication year - 1994
Publication title -
mathematics of operations research
Language(s) - English
Resource type - Journals
SCImago Journal Rank - 1.619
H-Index - 83
eISSN - 1526-5471
pISSN - 0364-765X
DOI - 10.1287/moor.19.1.152
Subject(s) - markov decision process , mathematics , mathematical optimization , markov chain , markov process , axiom , markov model , markov property , partially observable markov decision process , set (abstract data type) , mathematical economics , computer science , statistics , geometry , programming language
We consider a discrete time Markov Decision Process with infinite horizon. The criterion to be maximized is the sum of a number of standard discounted rewards, each with a different discount factor. Situations in which such criteria arise include modeling investments, production, modeling projects of different durations and systems with multiple criteria, and some axiomatic formulations of multi-attribute preference theory. We show that for this criterion for some positive e there need not exist an e-optimal randomized stationary strategy, even when the state and action sets are finite. However, e-optimal Markov nonrandomized strategies and optimal Markov strategies exist under weak conditions. We exhibit e-optimal Markov strategies which are stationary from some time onward. When both state and action spaces are finite, there exists an optimal Markov strategy with this property. We provide an explicit algorithm for the computation of such strategies and give a description of the set of optimal strategies.
Accelerating Research
Robert Robinson Avenue,
Oxford Science Park, Oxford
OX4 4GP, United Kingdom
Address
John Eccles HouseRobert Robinson Avenue,
Oxford Science Park, Oxford
OX4 4GP, United Kingdom