Introduction
Spiking neural networks (SNNs) are a well-established model of neural computation [1]. In contrast to conventional artificial neural networks (ANNs), neurons in an SNN communicate via stereotypical pulses—so-called spikes—and temporally integrate incoming information in their membrane potential. Since these are key features of biological neurons, SNNs are heavily used to model information processing in the brain [2], [3], [4]. Furthermore, SNNs are well-suited for implementation in neuromorphic hardware [5], leading to highly energy-efficient AI applications [6], [7], [8], [9].
For a long time, SNNs have been inferior to ANNs in terms of performance on standard pattern recognition tasks. However, a number of recent advances in SNN research have changed the picture, showing that SNNs can achieve performances similar to ANNs [10]. In particular, the use of surrogate gradients for SNN training [11], [12], [13] and the use of longer adaptation time constants in recurrent SNNs have been instrumental in this respect [14]. Nevertheless, SNNs still lack many capabilities of their biological counterparts—for some of which biologically implausible ANN solutions have been proposed.
Since computations in SNNs have—in contrast to computations in feed-forward ANNs—a strong temporal component, they have been proposed to be particularly suited for temporal computing tasks [14]. Here, it has turned out that the ability to retain information on several time scales is crucial. The most basic time constant in spiking neurons is the membrane time constant on the order of tens of milliseconds. In principle, arbitrary time constants can be realized by recurrent connections in recurrent SNNs. However, such recurrent retention of information is rather brittle and hard to learn. Instead, there were several suggestions to utilize longer time constants available in biological neuronal circuits such as short-term plasticity [15], [16] on the order of hundreds of milliseconds, and adaptation time constants of neurons [3], [14] on the order of seconds. The inclusion of such time constants has been shown to extend the computational capabilities of SNNs. However, typical cognitive tasks are frequently situated on a much slower time scale of minutes or longer. For example, when we watch a movie, we have to rapidly memorize facts to follow the story and draw conclusions as the narrative evolves. For such tasks, time constants on the order of seconds are insufficient.
Here, we consider Hebbian synaptic plasticity [17] as a mechanism to extend the range of time constants and therefore the computational capabilities of SNNs. Hebbian synaptic plasticity is abundant in both the neocortex and hippocampus [18], [19], [20]. While many forms, in particular, in sensory cortical areas, are believed to shape processing on a very slow developmental scale, there is also evidence for rapid plasticity that can in principle be utilized for online processing on the behavioral time scale, most prominently in the hippocampus [21], [22].
We present a novel memory-augmented SNN model that is equipped with a hetero-associative memory subject to Hebbian plasticity. Previous work has so far explored the concept of rapidly changing weights for memory only in the context of conventional ANNs [23], [24], [25], [26]. It has been shown that the integration of some type of memory into neural networks can strongly enrich their computational capabilities, which was previously achieved with rather unbiological types of memory components [27], [28], [29], [30], [31], [32], [33], [34] or with heavy weight sharing [35]. In contrast, our model is based on a novel associative memory component implemented by biological Hebbian plasticity for SNNs.
We experimentally show that our novel SNN model enriched by Hebbian plasticity outperforms state-of-the-art deep-learning mechanisms of long short-term memory (LSTM) networks [36], [37] and the long short-term memory spiking neural networks (LSNNs) [3], [14] in a sequential pattern-memorization task, as well as demonstrate superior out-of-distribution generalization capabilities compared to these models. The contemporary exceptional performance of standard deep-learning mechanisms strictly relies on the availability of a large number of training examples, whereas humans are capable of learning new tasks based on a single exposure (one-shot learning) [38]. We show that our memory-equipped SNN model provides a novel SNN-based solution to this problem and demonstrate that it can be successfully applied to one-shot learning and classification of handwritten characters, improving over previous SNN models [39], [40]. We also demonstrate the capability of our model to learn associations for audio-to-image synthesis from spoken and handwritten digits. Our SNN model enriched by Hebbian plasticity further presents a novel solution to a variety of cognitive question-answering tasks from a standard benchmark [27], achieving comparable performance to both memory-augmented ANN [41] and SNN-based [9] state-of-the-art solutions to this problem with a simpler architecture. In a final application scenario, we demonstrate that our model can learn from rewards on an episodic reinforcement learning task and attain a near-optimal strategy on a memory-based card game.
In contrast to H-mem, a previously proposed nonspiking model that utilizes Hebbian plasticity [41], our SNN model is well-suited for implementation in energy-efficient neuromorphic hardware. To demonstrate the potential efficiency advantages of such an implementation, we analyze its energy efficiency and compare it to H-Mem.
In summary, our results show that Hebbian plasticity enhances the computational capabilities of SNNs in several directions including out-of-distribution generalization, one-shot learning, cross-modal generative association, answering questions about stories of various types, and memory-dependent reinforcement learning. This suggests that Hebbian plasticity is a central component of information processing in the brain and artificial spike-based computing systems. Since local Hebbian plasticity can easily be implemented in neuromorphic hardware, this also suggests that powerful cognitive neuromorphic systems can be built based on this principle.
Method
We consider networks of standard leaky integrate-and-fire (LIF) neurons modeled in discrete time steps \begin{equation*} V_{j}(t + \Delta t) = \alpha V_{j}(t) + (1 - \alpha) I_{j}(t) - \vartheta z_{j}(t) \tag{1}\end{equation*}
The considered network model is shown in Fig. 1. At the core of the network is a heteroassociative memory [42], that is, a single-layer feed-forward SNN. Here, spiking neurons
Schematic of the SNN model. Inputs
In our model, neurons in the key layer (key neurons) receive input from two spiking neuron populations of size
Our model takes a sequence of input tokens
Consider a fact input \begin{equation*} I^{\mathrm {key}}_{j}(t) = \sum _{i} W^{\mathrm {s,key}}_{ji} z_{i}^{\mathrm {s,enc}}(t) \tag{2}\end{equation*}
\begin{equation*} I^{\mathrm {value}}_{k}(t) = \sum _{i} W^{\mathrm {s,value}}_{ki} z_{i}^{\mathrm {s,enc}}(t) + c \sum _{j} W^{\mathrm {assoc}}_{kj}(t) z_{j}^{\mathrm {key}}(t) \quad \tag{3}\end{equation*}
\begin{align*}&\hspace {-0.5pc}\Delta W_{kj}^{\mathrm {assoc}}\left ({t}\right) =\gamma _{+} \left ({w^{\mathrm {max}} - W_{kj}^{\mathrm {assoc}}\left ({t}\right)}\right)\,\kappa _{k}^{\mathrm {value}}\left ({t}\right) \kappa _{j}^{\mathrm {key}}\left ({t}\right) \\&- \gamma _{-} W_{kj}^{\mathrm {assoc}}\left ({t}\right) \kappa _{j}^{\mathrm {key}}\left ({t}\right)^{2} \tag{4}\end{align*}
Now, consider a query input \begin{align*}&\hspace {-0.5pc}I^{\mathrm {key}}_{j}(t) = \sum _{i \le d} W^{\mathrm {r,key}}_{ji} z_{i}^{\mathrm {r,enc}}(t) \\&+ \sum _{d \le k \le d+l} W^{\mathrm {r,key}}_{jk} z_{k-d}^{\mathrm {value}}(t-d_{\mathrm {feedback}}) \tag{5}\end{align*}
\begin{equation*} I^{\mathrm {value}}_{k}(t) = \sum _{j} W_{kj}^{\mathrm {assoc}}(t) z_{j}^{\mathrm {key}}(t) \tag{6}\end{equation*}
Finally, the value neurons project to an output network. The architecture of the output network depends on the task at hand. In the simplest case, the network output was determined by taking the sum of \begin{equation*} \hat {o}_{j} = \sum _{k} W^{\mathrm {out}}_{jk} \sum _{t^{\prime } = 0}^{\tau _{\mathrm {read}}} z_{k}^{\mathrm {value}}(M\tau _{\mathrm {sim}} - t^{\prime}). \tag{7}\end{equation*}
During training for all tasks, the weights
Results
A. Memorizing Associations
We first tested the ability of our model to one-shot memorize associations and to use these associations later when needed. Here, we conducted experiments on a task that requires to form of associations between random continuous-valued vectors and integer labels that were sequentially presented to the network.
In each instance of this task, we randomly drew
Association task and out-of-distribution generalization. (a) Schematic of the network model (left), the network input for an input sequence of length
The model was trained for 4250 iterations using backpropagation through time (BPTT) [3], [47]. We used two dense layers as input encoders in this task. Each layer consisted of 80 LIF neurons. One layer was used to encode the random input vectors and another layer was used to encode the integer labels. Inputs were applied to these layers for 100 ms each, giving rise to spike trains
In Fig. 2(b), we compare the performance of our model for various sequence lengths (number of vector-label pairs) to the standard generic artificial and spiking recurrent network models: the standard LSTM network [36] and the LSNN [3], [14]. The LSTM network consisted of 100 LSTM units. The LSNN consisted of 200 regular spiking and 200 adaptive neurons. For both models, we used the same architecture for the input encoder and the output layer as in our model. For short sequences, the performance of the LSTM network and the LSNN is comparable to our model. While the accuracy of the LSTM and LSNN drastically drops at some point, the accuracy of our model stays above 90% for sequences containing up to 50 vector-label pairs. The training of the network included a firing rate regularization term (see Supplementary), which resulted in a rather low average firing rate of 13 Hz per neuron. Hence, each neuron fired on average 13 spikes/s or approximately one spike for each 100 ms input token.
We next asked whether the network can generalize from a given sequence length to different sequence lengths. To this end, we trained it with a single sequence length of
We further asked whether the network is also able to generalize to different input firing rates. Similar to our previous approach, we trained the network with a single sequence length of 2, utilizing labels randomly selected from the set
B. One-Shot Learning
While standard deep-learning approaches need large numbers of training examples during training, humans can learn new concepts based on a single exposure (one-shot learning) [48]. A large number of few-shot learning approaches have been proposed using artificial neural networks [49], [50], but biologically plausible SNN models are very rare [39]. We wondered whether Hebbian plasticity could endow SNNs with one-shot learning capabilities. To that end, we applied our model to the problem of one-shot classification on the Omniglot [51] dataset. Omniglot consists of 1623 handwritten characters from 50 different alphabets. Each character is considered as one class. There are 20 examples of each class, each hand drawn by a different person. Following [52], we resized all the images to
We used a convolutional neural network (CNN) as the input encoder for the Omniglot image. This CNN was first pretrained alone on the training set using the prototypical loss [49] and then converted into a spiking CNN by using a threshold-balancing algorithm [53], [54], which is known to be the state-of-the-art approach in facilitating learning for deep convolutional SNNs [55], [56], [57] (see Section Converting Pre-Trained CNNs in the Supplementary for details on the conversion algorithm).
In each instance of the one-shot learning task, we randomly drew five different Omniglot classes. We then generated an input sequence as follows: From each of these five classes, we randomly drew a character instance and generated a random sequence out of them (i.e., the sequence is a random permutation of the five instances). This sequence was shown to the network together with a sequence of labels from 1 to 5 (again, a random permutation). Hence, each input character was paired with a unique label [see Fig. 3(a) (right)]. After all those image-label pairs have been presented to the network, it receives a query image. The query image showed another randomly drawn sample from one of these five Omniglot classes. The network was required to output the label that appeared together with an image of the same class as the query image (one-shot five-way classification).
Omniglot one-shot task and a visualization of the embeddings learned by the encoder CNN in this task. (a) Schematic of the network model (left) and the network input (right). Five images from the Omniglot dataset along with an associated label from 1 to 5 are presented sequentially as facts (
We trained our model for 200 epochs with 200 iterations/ epochs on the training set. In each iteration, we randomly drew a batch of 256 input sequences. During this training, the whole model was optimized, including the pretrained encoder CNN. We treated the grayscale values of an Omniglot image as a constant input current, applied for 100 ms to 784 LIF neurons, to produce input spikes to the CNN. The final layer of the CNN consisted of 64 LIF neurons. A single dense layer consisting also of 64 LIF neurons was used to encode the integer labels as spike trains with a duration of 100 ms per label (see Section Model and Training Details in the Supplementary for more details to the model and the training setup).
The rationale for this network architecture is that, given a suitable generalizing representation of the character, the Hebbian weight matrix can easily associate characters to labels, thus performing one-shot memorization of previously unseen classes to arbitrary labels. In biology, suitable representations could emerge from evolutionary optimized networks potentially fine-tuned by unsupervised plasticity processes. In our setup, these representations are provided by the CNN encoder, where the prototypical loss ensures that similar inputs are mapped to similar representations [49]. The network model can thus be seen as a spiking implementation of a prototypical network. However, the biologically unrealistic nearest-neighbor algorithm used to determine the output of the latter is replaced here by a simple heteroassociative memory. In Fig. 3(b), we show a sample
In a previous work [39], an SNN achieved an accuracy of 83.8% on a variant of the standard Omniglot one-shot learning task considered here. Instead of Hebbian plasticity, the model relied on a more elaborate three-factor learning rule for rapid learning. This model used a convolutional input encoder consisting of nonspiking neurons with batch normalization, which was trained end-to-end together with the SNN with BPTT. Recently, [40] presented a spiking CNN-based implementation of SiameseNets [58] and model-agnostic meta-learning (MAML) [50], which are other classical few-shot learning algorithms. Their spiking CNN implementations with a similar direct input encoding achieved 89.1% (spiking SiameseNets) and 91.5% (spiking MAML) one-shot five-way accuracies on Omniglot. Our model’s accuracy of 92.2% is a slight improvement over these results. Note, however, that these models were specifically designed for few-shot learning, while this is only one application of our versatile SNN model with Hebbian plasticity.
C. Cross-Modal Associations
Humans can imagine features of previously encountered stimuli. For example, when you hear the name of a person, you can imagine a mental image of their face. Here, in contrast to the associations considered above, not just classes are associated but (approximate) mental images. We, therefore, asked whether Hebbian plasticity can enable SNNs to perform such cross-modal associations. We trained our model in an autoencoder-like fashion. We used the FSDD [59] and the MNIST [60] datasets in this task. FSDD is an audio/speech dataset consisting of recordings of spoken digits. The dataset contains 3000 recordings of 6 speakers (50 of each digit per speaker).
In each instance of the task, we randomly drew three unique digits between 0 and 9. For each digit, we then generated a pair containing a randomly drawn instance of this digit as an audio file from the FSDD dataset, and a randomly drawn image of the same digit from the MNIST dataset [Fig. 4(a) (right)]. After these audio–image pairs have been presented to the network, it received an additional audio query. The audio query was another randomly drawn instance from the FSDD dataset of one of the previously presented digits. The network was required to generate the image of the handwritten digit that appeared together with the spoken digit of the same class as the audio query.
Audio to image synthesis task and examples of generated images. (a) Schematic of the network model (left) and the network input (right). Three audio–image pairs are presented sequentially as facts. Input pairs (
We used one CNN to encode the audio input (more specifically the mel-frequency cepstrum coefficients (MFCCs); see section Model and Training Details in the Supplementary) and one CNN to encode the MNIST images. The CNNs were pre-trained on FSDD/MNIST classification tasks, respectively, and the pretrained models were converted into spiking CNNs by using the threshold-balancing algorithm [53], [54] (see Section Converting Pre-Trained CNNs in the Supplementary for details to the conversion algorithm). We removed the final classification layers of the CNNs and used the penultimate layer consisting of 64 LIF neurons as the encoding of the input stimuli. The spiking CNN was then fine-tuned during end-to-end training (see Section Model and Training Details in the Supplementary for more details on the model and the training setup). Following the value layer of the model, the image reconstruction was produced by a two-layer fully-connected network with 256 and 784 LIF neurons, respectively.
In Fig. 4(b), we show example MNIST images from the test set and the corresponding images that were reconstructed by the network. One can see that not just a typical image for the digit class was imagined by the network, but rather an image that is very similar to the image presented previously with the audio cue. This shows that the network did not just memorize the digit class in its associative memory, but rather features that benefit the reconstruction of this specific sample. To quantify the reconstruction performance of our model, we computed the mean squared difference (MSD) between the image produced by the network and all MNIST images in an input sequence. The MSD was 0.03± 0.02 (mean ± standard deviation; median was 0.02 with a lower and upper quartile of 0.01 and 0.03, respectively) between the reconstructed image and the target image, and 0.1 ± 0.04 between the reconstructed image and the two other MNIST images in the input sequence (statistics are over 1000 examples in the test set). Again, the network operated in a sparse activity regime with an average firing rate of 10 Hz.
D. Question Answering
The bAbI dataset [27] is a standard benchmark for cognitive architectures with memory. The dataset contains 20 different types of synthetic question-answering (QA) tasks, designed to test a variety of reasoning abilities on stories. Each of these tasks consists of a sequence of sentences followed by a question whose answer is typically a single word (in a few tasks, answers are multiple words; see Table S2 in the Supplementary for example stories and questions). We provided the answer to the model as supervision during training, and it had to predict it at test time on a separate test set. The performance of the model was measured using the average error on the test dataset over all tasks and the number of failed tasks (according to the convention of [27], a model had failed to solve a task if the test error was above 5% for that task).
Each instance of a task consists of a sequence of
We used a dense layer consisting of 80 LIF neurons as input encoder in this task. We found it helpful to let the model choose for itself which type of sentence encoding to use. We, therefore, used a learned encoding (see Section Model and Training Details in the Supplementary and [35]). Each sentence was encoded as a spike train with a duration of 100 ms. Similar to previous work [28], [35], we performed three independent runs with different random initializations and reported the results of the model with the highest validation accuracy in these runs. The average firing rate per neuron estimated over all bAbI tasks was 12 Hz, which corresponds to approximately one spike for each 100 ms input token.
In Table II, we compare our model to the spiking RelNet [9] and to the H-Mem model [41], a nonspiking memory network model. Similar to the feedback-loop from the value layer to the input of the key layer in our model, the H-Mem model can utilize several memory accesses conditioned on previous memory recalls. The results of H-Mem were reported for a single memory access (1-hop) and three memory accesses (3-hop). It turned out that multiple memory hops were necessary to solve some of the bAbI tasks. Similarly, we found that in our model, an instantaneous feedback loop (with a delay
E. Reinforcement Learning
While supervisory signals are arguably scarce in nature, it is well-established that animals learn from rewards [61]. To test whether memory can also serve SNNs in the reward-based setting, we evaluated our model on an episodic reinforcement learning task. The task is based on the popular children’s game Concentration. The game Concentration requires good memorization skills and is played with a deck of
Here, we consider a one-player solitaire version of the game [Fig. 5(a)]. In this version of the game, the objective is to find all matching pairs with as few card flips as possible. The cards are arranged on a 1-D grid of cells, each of which may be empty or may contain one card. The grid is just large enough to hold all of the
SNN learns to play the game Concentration from rewards. (a) Example game moves for a Concentration game with four cards played as a solitaire. The objective of the game is to turn over pairs of matching cards with as few card flips as possible. Shown is—for each time step
Instead of using images, we define each card face to be a 10-D random continuous-valued vector. The agent’s observation vector
The performance was evaluated in terms of the number of flips performed until all matching pairs had been removed from the grid. Agents were trained with proximal policy optimization (PPO) [62] in actor–critic style using 64 asynchronous vectorized environments (see Section Model and Training Details in the Supplementary for details of the model and the training setup). We evaluated our model on a deck of four cards and a deck of six cards. The evolution of the number of cards flips the agent takes to finish a game over the number of games during training is shown in Fig. 5(b). After training, we evaluated the agents on 1000 games and recorded the number of card flips the agent takes to finish each game. Fig. 5(c) shows the histogram of the number of cards flips in this evaluation for a random agent, our agent, and an agent that has perfect memory and that follows an optimal strategy. If an agent has no memory at all and plays by simply flipping cards at random, then the expected number of flips the agent takes to finish a game with
F. Ablation Studies
The different weight matrices in our network serve different purposes (see Fig. 1). The matrix
We first generated independent random weight matrices for
To accomplish rapid associations during inference, it is necessary that Hebbian plasticity can elevate membrane potentials of neurons above the threshold in the value layer within the time of an input token presentation (100 ms in our setup). This means that neuron thresholds should be relatively small compared to the maximum weights of the association matrix. In all our simulations, we used a threshold of
Efficiency of Neuromorphic Implementation
SNNs can be implemented highly efficiently in analog neuromorphic hardware [64], [65], [66]. However, recently energy-efficient digital implementations have been introduced [5], [67]. The main advantage of digital SNNs over digital ANNs is the drastic reduction of multiply-accumulate (MAC) operations. Since spikes are binary, for every input spike, the weight has to be added to the membrane potential of the neuron, which is one accumulate operation (AC). Since MAC operations are much more expensive than AC operations (
As an illustrative example, we consider the association task from Section III-A. The core network without the input encoder and the readout, consists of
Discussion
Local Hebbian plasticity is a key ingredient of biological neural systems. While only local Hebbian plasticity is needed during inference in our model, it was trained with BPTT. Since BPTT is biologically implausible, we cannot claim that our network is a model for how functionality could be learned by an organism. Instead, our results provide an existence proof for powerful memory-enhanced SNNs. One can speculate that brain networks were shaped by evolution to make use of Hebbian plasticity processes. In this sense, BPTT can be seen as a replacement for evolutionary processes. For example, the brain might have evolved networks for one-shot learning (Fig. 3) that are particularly tuned to behaviorally relevant stimuli. In addition to evolutionary optimization, local approximations of BPTT such as the recently introduced e-prop algorithm [70] could then further shape the evolved circuits for specific functionality.
The integration of synaptic plasticity for inference in artificial neural networks was used in [23] and [24] and adopted recently [25], [26]. In the latter, input representations were bound to labels with Hebbian plasticity. Memory-augmented neural networks use explicit memory modules which are a differentiable version of a digital memory [27], [28], [29], [30], [31], [32], [33]. Our model utilizes biological Hebbian plasticity instead. The use of fast Hebbian plasticity in SNNs was studied in [71] and [72]. These models implement a type of variable binding and can explain certain aspects of higher-level cognition in the brain. They were, however, not applied to complex cognitive tasks. In [73], training of networks with synaptic plasticity was explored, but there the parameters of the plasticity rule were optimized instead of the surrounding control networks. Other uses of Hebbian plasticity for network computations include the unsupervised pretraining followed by supervised training of the output layer [74] and the derivation of local plasticity rules for unsupervised learning [75].
The inclusion of Hebbian plasticity can be viewed as the introduction of another long time constant in the network dynamics. Previous work has shown that longer time constants can significantly improve the temporal computing capabilities of SNNs. In this direction, short-term synaptic plasticity [15], [16] and neuronal adaptation [14] have been exploited. One-shot learning of SNNs has been studied in [39]. Instead of Hebbian plasticity, this model relied on a more elaborate three-factor learning rule. Another SNN model, the spiking RelNet, was tested on the bAbI task set [9]. We have compared this model to ours in Table II. The architecture of this model is quite different from our proposal. As a spiking implementation of relational networks, it is rather complex and makes heavy use of weight sharing. Both these models perform well on similar tasks of the bAbI task set, but interestingly, there are some differences. For example, our model solves task 2 “two supporting facts” on which the spiking RelNet fails. This might be due to the possibility of multiple memory accesses through the feedback loop in our model. On the other hand, our model fails at task 17 “positional reasoning” which is solved by the spiking RelNet. We suspect that this is due to the more complex network structure of the spiking RelNet. To the best of our knowledge, no previous spiking (or artificial) neural network model is performing well on both, one-shot learning and bAbI tasks, as well as on the other tasks we presented.
Our results on bAbI tasks show that a synaptic delay in the feedback connections from the value layer to the key layer is essential for good performance in some tasks. While we used fixed delays of 1 and 30 ms in our simulations, an intriguing option is to optimize this delay during training. In general, optimizing delays together with weights in the network [76] could lead to more efficient networks that can better utilize spike timing information. This option could be investigated in future work.
Our discussion of efficient network implementation in Section IV did not include Hebbian updates of the association weights. Since here, multiplication is performed at every time step, the vanilla spiking model is not efficient. To optimize the performance of the Hebbian updates, one could perform only one update per input token based on the product of the number of pre- and postsynaptic spikes during the token presentation. This would use computational resources comparable to the nonspiking case. Even more efficient would be the use of analog memories such as memristors [77], for which efficient implementations of Hebbian plasticity have been proposed [78].
Conclusion
We have presented a novel SNN model that integrates Hebbian plasticity in its network dynamics. We found that this memory enrichment renders SNNs surprisingly flexible. In particular, our simulations show that Hebbian plasticity endows SNNs with the capability for one-shot learning, cross-modal pattern association, language processing, and memory-dependent reward-based learning. Hebbian plasticity is spatially and temporally local, i.e., the synaptic weight change depends only on a filtered version of the pre- and postsynaptic spikes and on the current weight value. This is a very desirable feature of any plasticity rule, as it can easily be implemented both in biological synaptic connections and in neuromorphic hardware. In fact, current neuromorphic designs support this type of plasticity [5]. Hence, our results indicate that Hebbian plasticity can serve as a fundamental building block in cognitive architectures based on energy-efficient neuromorphic hardware.
ACKNOWLEDGMENT
The authors would like to thank Wolfgang Maass and Arjun Rao for initial discussions.