학술논문

Learning policies for abstract state spaces
Document Type
Conference
Source
2005 IEEE International Conference on Systems, Man and Cybernetics System, Man and Cybernetics Systems, Man and Cybernetics, 2005 IEEE International Conference on. 4:3179-3184 Vol. 4 2005
Subject
Robotics and Control Systems
Computing and Processing
Components, Circuits, Devices and Systems
State-space methods
Dynamic programming
Multidimensional systems
Partitioning algorithms
Cost function
Iterative algorithms
Language
ISSN
1062-922X
Abstract
Applying Q-learning to multidimensional, real-valued state spaces is time-consuming in most cases. In this article, we deal with the assumption that a coarse partition of the state space is sufficient for learning good or even optimal policies. An algorithm is presented which constructs proper policies for abstract state spaces using an incremental procedure without approximating a Q-function. By combining an approach similar to dynamic programming and a search for policies, we can speed up the learning process. To provide empirical evidence, we use a cart-pole system. Experiments were conducted for a simulated environment as well as for a real plant.