We set out to create a memristive synapse that can express the different plasticity rules, such as long-term and short-term plasticity and their combinations using commercial PCM — at the nanoscale. We achieved this by using the combination of the non-volatility (from amorphous-crystalline phase transitions) and volatility (from electrostatics change) in PCMs.
Phase change materials have been independently researched for memory and transistor applications, but their combination for neuromorphic computing has not been previously explored. Our demonstration shows how the non-volatility of the devices can enable long-term plasticity while the volatility provides the short-term plasticity and their combination allows for other mixed-plasticity computations, much like in the mammalian brain.
We began testing our idea toward the end of 2019, starting by using the volatility property in the devices for a different application: compensating for the non-idealities in phase change computational memories. Our team fondly remembers chatting over coffee, discussing the first experimental results that would allow us to take the next step of building the synthetic synapses we originally imagined. With some learning and trial and error during the pandemic, by early 2021 we had our first set of results and things worked out the way we had imagined.
To test the utility of our device, we implemented a curated form of an algorithm for sequential learning which our co-author, Timolean Moraitis, was developing to use spiking networks for learning dynamically changing environments. From this emerged our implementation of a short-term spike-timing-dependent plasticity framework that allowed us to demonstrate a biologically inspired form of sequentially learning on artificial hardware. Instead of the playful cat that we mentioned earlier, we showed how machine vision can recognize the moving frames of a boy and a girl.
Later, we expanded the concept to show the emulation of other biological processes for useful computations. Crucially, we showed how the homeostatic phenomena in the brain could provide a means to construct efficient artificial hardware for solving difficult combinatorial optimization problems. We achieved this by constructing stochastic Hopfield neural networks, where the noise mechanisms at the synaptic junction provide efficiency gains to computational algorithms.
Our results are more exploratory than actual system-level demonstrations. In the latest paper, we propose a new synaptic concept that does more than a fixed synaptic weighing operation, as is the case in modern artificial neural networks. While we plan to take this approach further, we believe our existing proof-of-principle results already elucidate significant scientific interest in the broader areas of neuromorphic engineering — in computing and understanding the brain better from its more faithful emulations.
Our devices are based on a well-researched technology and are easy to fabricate and operate. The key challenge we see is the at-scale implementation, stringing together our computational primitive and other hardware blocks. Building a full-fledged computational core with our devices will require a rethinking of the design and implementation of peripheral circuitries.
Currently, these are standard to conventional memristive computational hardware: we require additional control terminals and resources for our devices to function. There are some ideas in place, such as redefining the layouts, as well as restructuring the basic device design, which we are currently exploring.
Our current results show how relevant mixed-plasticity neural computations can prove to be in neuromorphic engineering. The demonstration of sequential learning can allow neural networks to recognize and classify objects more efficiently. This not only makes, for example, visual cognition more human-like but also provides significant savings on the expensive training processes.
Our illustration of a Hopfield neural network allows us to solve difficult optimization problems. We show an example of max cut, which is a graph coloring problem and has utility in applications such as chip design. Other applications include problems like flight scheduling, internet packets routing, and more.
This research was partially funded by European Union’s Horizon 2020 research and innovation program (Fun-COMP project, Grant Number 780848) and the European Research Council through the European Union’s Horizon 2020 Research and Innovation Program under grant number 682675.
Sarwat, S.G., Kersting, B., Moraitis, T. et al. Phase-change memtransistive synapses for mixed-plasticity neural computations. Nat. Nanotechnol. (2022). ↩