Abstract
The delta rule and its generalisations is one of the most practically useful learning algorithms for connectionist systems. (See [5] and, e.g. [1]). It is often stated that the method is gradient descent for the least squares error. However, neither part of this statement is true for finitely large learning rates. The purpose of this paper is to employ certain mathematical tools, largely borrowed from numerical analysis, to provide a rigourous framework for discussing the behaviour of learning algorithms. To this end, a more or less complete analysis of the original linear version of the delta rule is presented. It is shown that when the rule is applied by repetitively cycling through a fixed epoch of patterns, updating the weights after each pattern, the algorithm generates a limit cycle. The least squares error of this limit cycle appoaches that of the true minimum quadratically as the learning rate tends to zero. The algorithm is convergent and numerically stable subject only to a simple normalisation condition. By contrast, if the weights are updated after the complete epoch of patterns has been presented, the iteration has a fixed point which is the true least squares minimum. However the algorithm may have very bad numerical stability and convergence properties, even for problems which are “good” from the point of view of learning. This simple linear case is of limited practical use, but heuristic and numerical evidence suggests that the analysis does give insight into the behaviour of the method for more useful cases such as back propagation networks. Current work is directed to a rigorous justification of this. It is the author’s belief that the methods can be extended to other learning paradigms.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Preview
Unable to display preview. Download preview PDF.
Similar content being viewed by others
References
M.R.Devos and G.A.Orban: “Self-adapting back propagation”, in the proceedings of Neuro-Nimes `88, Nimes, France, November 1988, from EC2, 269–287 rue de la Garenne, 92000, Nanterre, France.
S.Ellacott: “Some working papers on the delta rule”, ITRI technical report no.79, 1989. (Address as above.)
E.Isaacson and H.B.Keller: “Analysis of numerical methods”, Wiley, 1966.
R.D.Milne: “Applied functional analysis: an treatment”, Pitman, 1980. (ISBN 0–273–08404–6)
D.E.Rumelhart and J.L.McClelland: “Parallel and distributed processing: explorations in the microstructure of cognition”, vol.1, MIT, 1986. (ISBN 0–262–181270–7)
Author information
Authors and Affiliations
Rights and permissions
Copyright information
© 1990 Springer Science+Business Media Dordrecht
About this chapter
Cite this chapter
Ellacott, S.W. (1990). An Analysis of the Delta Rule. In: International Neural Network Conference. Springer, Dordrecht. https://doi.org/10.1007/978-94-009-0643-3_145
Download citation
DOI: https://doi.org/10.1007/978-94-009-0643-3_145
Publisher Name: Springer, Dordrecht
Print ISBN: 978-0-7923-0831-7
Online ISBN: 978-94-009-0643-3
eBook Packages: Springer Book Archive