Abstract
This paper presents a new adaptive segmentation of continuous state space based on vector quantization algorithm such as Linde–Buzo–Gray for high-dimensional continuous state spaces. The objective of adaptive state space partitioning is to develop the efficiency of learning reward values with an accumulation of state transition vector in a single-agent environment. We constructed our single-agent model in continuous state and discrete actions spaces using Q-learning function. Moreover, the study of the resulting state space partition reveals a Voronoi tessellation. In addition, the experimental results show that this proposed method can partition the continuous state space appropriately into Voronoi regions according to not only the number of actions, but also achieve a good performance of reward-based learning tasks compared with other approaches such as square partition lattice on discrete state space.
Similar content being viewed by others
References
Sutton RS, Barto AG (1998) Reinforcement learning an introduction. MIT Press, Cambridge
Watkins CJCH, Dayan P (1992) Technical notes: Q-learning. Mach Learn 8:279–292
Patane G, Russo M (2001) The enhanced LBG algorithm. In: Proceeding of neural networks, pp 1219–1237
Shen F, Hasegawa O (2006) An adaptive incremental LBG for vector quantization. Neural Netw 19:694–704
Author information
Authors and Affiliations
Corresponding author
About this article
Cite this article
Fuchida, T., Aung, K.T. A proposition of adaptive state space partition in reinforcement learning with Voronoi tessellation. Artif Life Robotics 18, 172–177 (2013). https://doi.org/10.1007/s10015-013-0125-x
Received:
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s10015-013-0125-x