Applying Q-Learning to Non-Markovian Environments
Proceedings of the First International Conference on Agents and Artificial Intelligence (ICAART 2009) 2009
Jurijs Čižovs, Arkādijs Borisovs

This paper considers the problem of intelligent agent functioning in non-Markovian environments. We advice to divide the problem into two subproblems: just finding non-Markovian states in the environment and building an internal representation of original environment by the agent. The internal representation is free from non Markovian states because insufficient number of additional dynamically created states and transitions are provided. Then, the obtained environment might be used in classical reinforcement learning algorithms (like SARSA(λ)) which guarantee the convergence by Bellman equation. A great difficulty is to recognize different “copies” of the same states. The paper contains a theoretical introduction, ideas and problem description, and, finally, an illustration of results and conclusions.


Atslēgas vārdi
Reinforcement learning, non-Markovian deterministic environments, intelligent agents, agent control

Čižovs, J., Borisovs, A. Applying Q-Learning to Non-Markovian Environments. No: Proceedings of the First International Conference on Agents and Artificial Intelligence (ICAART 2009), Portugāle, Porto, 19.-21. janvāris, 2009. Porto: Institute for Systems and Technologies of Information, Control and Communication, 2009, 306.-311.lpp. ISBN 9783642118197.

Publikācijas valoda
English (en)
RTU Zinātniskā bibliotēka.
E-pasts: uzzinas@rtu.lv; Tālr: +371 28399196