15th Triennial World Congress of the International Federation of Automatic Control
  Barcelona, 21–26 July 2002 
ADAPTIVE VECTOR QUANTIZATION FOR REINFORCEMENT LEARNING
H. Y. K. Lau, K. L. Mak and I. S. K. Lee
Department of Industrial and Manufacturing Systems Engineering
The University of Hong Kong, Pokfulam Road, Hong Kong.

Dynamic programming methods are capable of solving reinforcement learning problems, in which an agent must improve its behavior through trial-and-error interactions with a dynamic environment. However, these computational algorithms suffer from the curse of dimensionality (Bellman, 1957) that the number of computational operations increases exponentially with the cardinality of the state space. In practice, this usually results in a very long training time and applications in continuous domain are far from trivial. In order to ease this problem, we propose the use of vector quantization to adaptively partition the state space based on the recent estimate of the action-value function. In particular, this state-space partitioning operation is performed incrementally to reflect the experience accumulated by the agent as it explores the underlying environment.
Keywords: Learning algorithm, intelligent control, vector quantization, robot navigation, automated guided vehicles

E-mail: hyklau@hku.hk
Session slot T-Fr-A21: Posters of Learning, Stochastic, Fuzzy and Nerural Systems/Area code 3b : Adaptive Control and Tuning