|
This article is cited in 1 scientific paper (total in 1 paper)
Finding control policy for one discrete-time Markov chain on $[0, 1]$ with a given invariant measure
M. G. Konovalova, R. V. Razumchikab a Institute of Informatics Problems, Federal Research Center “Computer Science and Control” of the Russian Academy of Sciences, 44-2 Vavilov Str., Moscow 119333, Russian Federation
b Peoples' Friendship University of Russia (RUDN University), 6 Miklukho-Maklaya Str., Moscow 117198, Russian
Federation
Abstract:
A discrete-time Markov chain on the interval $[0,1]$ with two possible transitions (left or right) at each
step has been considerred. The probability of transition towards $0$ (and towards $1$) is a function of the current value
of the chain. Having chosen the direction, the chain moves to the randomly chosen point from the appropriate
interval. The authors assume that the transition probabilities depend on the current value of the chain only through a finite
number of real-valued numbers. Under this assumption, they seek the transition probabilities, which guarantee
the $L_2$ distance between the stationary density of the Markov chain and the given invariant measure on $[0,1]$
is minimal. Since there is no reward function in this problem, it does not fit in the
MDP (Markov decision process) framework. The authors follow the
sensitivity-based approach and propose the gradient- and simulation-based method for estimating the parameters of
the transition probabilities. Numerical results are presented which show the performance of the method for various
transition probabilities and invariant measures on $[0,1]$.
Keywords:
Markov chain; control; continuous state space; sensitivity-based approach; derivative estimation.
Received: 28.04.2018
Citation:
M. G. Konovalov, R. V. Razumchik, “Finding control policy for one discrete-time Markov chain on $[0, 1]$ with a given invariant measure”, Inform. Primen., 12:3 (2018), 2–13
Linking options:
https://www.mathnet.ru/eng/ia540 https://www.mathnet.ru/eng/ia/v12/i3/p2
|
Statistics & downloads: |
Abstract page: | 212 | Full-text PDF : | 71 | References: | 25 |
|