Skip to main content

Abstract

The delta rule and its generalisations is one of the most practically useful learning algorithms for connectionist systems. (See [5] and, e.g. [1]). It is often stated that the method is gradient descent for the least squares error. However, neither part of this statement is true for finitely large learning rates. The purpose of this paper is to employ certain mathematical tools, largely borrowed from numerical analysis, to provide a rigourous framework for discussing the behaviour of learning algorithms. To this end, a more or less complete analysis of the original linear version of the delta rule is presented. It is shown that when the rule is applied by repetitively cycling through a fixed epoch of patterns, updating the weights after each pattern, the algorithm generates a limit cycle. The least squares error of this limit cycle appoaches that of the true minimum quadratically as the learning rate tends to zero. The algorithm is convergent and numerically stable subject only to a simple normalisation condition. By contrast, if the weights are updated after the complete epoch of patterns has been presented, the iteration has a fixed point which is the true least squares minimum. However the algorithm may have very bad numerical stability and convergence properties, even for problems which are “good” from the point of view of learning. This simple linear case is of limited practical use, but heuristic and numerical evidence suggests that the analysis does give insight into the behaviour of the method for more useful cases such as back propagation networks. Current work is directed to a rigorous justification of this. It is the author’s belief that the methods can be extended to other learning paradigms.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 39.99
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 54.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

Preview

Unable to display preview. Download preview PDF.

Unable to display preview. Download preview PDF.

Similar content being viewed by others

References

  1. M.R.Devos and G.A.Orban: “Self-adapting back propagation”, in the proceedings of Neuro-Nimes `88, Nimes, France, November 1988, from EC2, 269–287 rue de la Garenne, 92000, Nanterre, France.

    Google Scholar 

  2. S.Ellacott: “Some working papers on the delta rule”, ITRI technical report no.79, 1989. (Address as above.)

    Google Scholar 

  3. E.Isaacson and H.B.Keller: “Analysis of numerical methods”, Wiley, 1966.

    Google Scholar 

  4. R.D.Milne: “Applied functional analysis: an treatment”, Pitman, 1980. (ISBN 0–273–08404–6)

    Google Scholar 

  5. D.E.Rumelhart and J.L.McClelland: “Parallel and distributed processing: explorations in the microstructure of cognition”, vol.1, MIT, 1986. (ISBN 0–262–181270–7)

    Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Rights and permissions

Reprints and permissions

Copyright information

© 1990 Springer Science+Business Media Dordrecht

About this chapter

Cite this chapter

Ellacott, S.W. (1990). An Analysis of the Delta Rule. In: International Neural Network Conference. Springer, Dordrecht. https://doi.org/10.1007/978-94-009-0643-3_145

Download citation

  • DOI: https://doi.org/10.1007/978-94-009-0643-3_145

  • Publisher Name: Springer, Dordrecht

  • Print ISBN: 978-0-7923-0831-7

  • Online ISBN: 978-94-009-0643-3

  • eBook Packages: Springer Book Archive

Publish with us

Policies and ethics