Efficient SpiNNaker simulation of a heteroassociative memory using the Neural Engineering Framework

The 2016 International Joint Conference on Neural Networks (IJCNN), 2016

James Knight, Aaron R. Voelker, Andrew Mundy, Chris Eliasmith, Steve Furber

Abstract

The biological brain is a highly plastic system within which the efficacy and structure of synaptic connections are constantly changing in response to internal and external stimuli. While numerous models of this plastic behavior exist at various levels of abstraction, how these mechanisms allow the brain to learn meaningful values is unclear. The Neural Engineering Framework (NEF) is a hypothesis about how large-scale neural systems represent values using populations of spiking neurons, and transform them using functions implemented by the synaptic weights between populations. By exploiting the fact that these connection weight matrices are factorable, we have recently shown that static NEF models can be simulated very efficiently using the SpiNNaker neuromorphic architecture. In this paper, we demonstrate how this approach can be extended to efficiently support both supervised and unsupervised learning rules designed to operate on these factored matrices. We then present a heteroassociative memory architecture built using these learning rules and prove that it is capable of learning a human-scale semantic network. Finally we demonstrate a 100 000 neuron version of this architecture running on the SpiNNaker simulator with a speed-up exceeding 150x when compared to the Nengo reference simulator.

Full text links

External link

Conference Proceedings

Publisher
IEEE
Booktitle
The 2016 International Joint Conference on Neural Networks (IJCNN)
Address
Vancouver, British Columbia
Month
07

Cite

Plain text

BibTeX