Abstract
Two approaches to control policy synthesis for unknown systems are investigated. An indirect approach is based on adaptive identification of a neural network model in the NAR form (nonlinear autoregresion model) followed by application of the dynamic programming to this model. A direct approach consists of Q-learning with the use of a lookup table. Both methods were applied to optimization of a stock portfolio problem and tested on Warsaw Stock Exchange data.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Preview
Unable to display preview. Download preview PDF.
References
D.P. Bertsekas and J. Tsitsiklis, Neuro-Dynamic Programming, Athena Scientific, Belmont, Mass., 1996.
D.P. Bertsekas, Dynamic Programming and Optimal Control, Athena Scientific, Belmont, Mass., 1995.
S. Haykin, Neural Networks - A Comprehensive Foundation Macmillan College Publishing Company, 1994.
K.S. Narendra, Neural Networks for Control: Theory and Practice, Proceedings of the IEEE, Vol. 84, No. 10, pp. 1385–1407, 1996.
C.J.C.H. Watkins and P. Dayan, Technical Note: Q Learning, Machine Learning, Vol. 8, pp. 279–492, 1992.
Author information
Authors and Affiliations
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2003 Springer-Verlag Berlin Heidelberg
About this paper
Cite this paper
Chrobak, J., Pacut, A. (2003). Dynamic Programming with NAR Model versus Q-learning — Case Study. In: Rutkowski, L., Kacprzyk, J. (eds) Neural Networks and Soft Computing. Advances in Soft Computing, vol 19. Physica, Heidelberg. https://doi.org/10.1007/978-3-7908-1902-1_113
Download citation
DOI: https://doi.org/10.1007/978-3-7908-1902-1_113
Publisher Name: Physica, Heidelberg
Print ISBN: 978-3-7908-0005-0
Online ISBN: 978-3-7908-1902-1
eBook Packages: Springer Book Archive