Concepedia

Publication | Open Access

Online Markov Decision Processes Under Bandit Feedback

102

Citations

23

References

2014

Year

Abstract

We consider online learning in finite stochastic Markovian environments where in each time step a new reward function is chosen by an oblivious adversary. The goal of the learning agent is to compete with the best stationary policy in hindsight in terms of the total reward received. Specifically, in each time step the agent observes the current state and the reward associated with the last transition, however, the agent does not observe the rewards associated with other state-action pairs. The agent is assumed to know the transition probabilities. The state of the art result for this setting is an algorithm with an expected regret of O(T <sup xmlns:mml="http://www.w3.org/1998/Math/MathML" xmlns:xlink="http://www.w3.org/1999/xlink">2/3</sup> lnT). In this paper, assuming that stationary policies mix uniformly fast, we show that after T time steps, the expected regret of this algorithm (more precisely, a slightly modified version thereof) is O(T <sup xmlns:mml="http://www.w3.org/1998/Math/MathML" xmlns:xlink="http://www.w3.org/1999/xlink">1/2</sup> lnT), giving the first rigorously proven, essentially tight regret bound for the problem.

References

YearCitations

1998

26.8K

2005

25.7K

1995

8.4K

2002

2.2K

2009

1.9K

2010

707

2004

442

2008

250

2008

246

2010

208

Page 1