Skip to main content

Simulating Biological-Inspired Spiking Neural Networks with OpenCL

  • Conference paper

Part of the book series: Lecture Notes in Computer Science ((LNTCS,volume 6352))

Abstract

The algorithms used for simulating biologically-inspired spiking neural networks (BIANN) often utilize functions which are computationally complex and have to model a large number of neurons - or even a much larger number of synapses in parallel. To use all available computing resources provided by a standard desktop PC is an opportunity to shorten the simulation time and extend the number of simulated neurons and their interconnections. OpenCL offers an open platform for heterogeneous computing to employ CPUs, GPUs, DSP or FPGAs in an uniform way. This paper introduces a handy simulation framework being sufficient to accelerate different kinds of neural networks with off-the-shelf hardware. To illustrate this, different large networks comprising a complex synaptic model in combination with a leaky Integrate-and-Fire neuron model are implemented as standard Matlab code and with OpenCL separately. In comparison to the Matlab model, OpenCL reaches a speedup of \(\backsim83\) on a quad-core processor and of \(\backsim1500\) on a GPU.

This is a preview of subscription content, log in via an institution.

Buying options

Chapter
USD   29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD   84.99
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD   109.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Learn about institutional subscriptions

Preview

Unable to display preview. Download preview PDF.

Unable to display preview. Download preview PDF.

References

  1. Blas, A.D., Jagota, A., Hughey, R.: Optimizing neural networks on simd parallel computers. Parallel Computing 31(1), 97–115 (2005)

    Article  Google Scholar 

  2. Bernhard, F., Keriven, R.: Spiking Neurons on GPUs, pp. 236–243. Springer, Heidelberg (2006)

    Google Scholar 

  3. Khan, M.M., Lester, D.R., Plana, L.A., Rast, A.D., Jin, X., Painkras, E., Furber, S.B.: Spinnaker: Mapping neural networks onto a massively-parallel chip multiprocessor. In: IJCNN, pp. 2849–2856 (2008)

    Google Scholar 

  4. Guzhva, A., Dolenko, S., Persiantsev, I.: Multifold acceleration of neural network computations using gpu. In: Alippi, C., Polycarpou, M., Panayiotou, C., Ellinas, G. (eds.) ICANN 2009. LNCS, vol. 5768, pp. 373–380. Springer, Heidelberg (2009)

    Chapter  Google Scholar 

  5. Izhikevich, E.M.: Simple model of spiking neurons. IEEE Trans. Neural Networks, 1569–1572 (2003)

    Google Scholar 

  6. El-Laithy, K., Bogdan, M.: Synchrony state generation in artificial neural networks with stochastic synapses. In: Alippi, C., Polycarpou, M., Panayiotou, C., Ellinas, G. (eds.) ICANN 2009. LNCS, vol. 5768, pp. 181–190. Springer, Heidelberg (2009)

    Chapter  Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2010 Springer-Verlag Berlin Heidelberg

About this paper

Cite this paper

Hoffmann, J., El-Laithy, K., Güttler, F., Bogdan, M. (2010). Simulating Biological-Inspired Spiking Neural Networks with OpenCL. In: Diamantaras, K., Duch, W., Iliadis, L.S. (eds) Artificial Neural Networks – ICANN 2010. ICANN 2010. Lecture Notes in Computer Science, vol 6352. Springer, Berlin, Heidelberg. https://doi.org/10.1007/978-3-642-15819-3_23

Download citation

  • DOI: https://doi.org/10.1007/978-3-642-15819-3_23

  • Publisher Name: Springer, Berlin, Heidelberg

  • Print ISBN: 978-3-642-15818-6

  • Online ISBN: 978-3-642-15819-3

  • eBook Packages: Computer ScienceComputer Science (R0)

Publish with us

Policies and ethics