Alain Vaucher, Philippe Schwaller, et al.
AMLD EPFL 2022
Kernel functions are vital ingredients of several machine learning (ML) algorithms but often incur substantial memory and computational costs. We introduce an approach to kernel approximation in ML algorithms suitable for mixed-signal analogue in-memory computing (AIMC) architectures. Analogue in-memory kernel approximation addresses the performance bottlenecks of conventional kernel-based methods by executing most operations in approximate kernel methods directly in memory. The IBM HERMES project chip, a state-of-the-art phase-change memory-based AIMC chip, is utilized for the hardware demonstration of kernel approximation. Experimental results show that our method maintains high accuracy, with less than a 1% drop in kernel-based ridge classification benchmarks and within 1% accuracy on the long-range arena benchmark for kernelized attention in transformer neural networks. Compared to traditional digital accelerators, our approach is estimated to deliver superior energy efficiency and lower power consumption. These findings highlight the potential of heterogeneous AIMC architectures to enhance the efficiency and scalability of ML applications.
Alain Vaucher, Philippe Schwaller, et al.
AMLD EPFL 2022
Masami Akamine, Jitendra Ajmera
IEICE Trans Inf Syst
Els van Herreweghen, Uta Wille
USENIX Workshop on Smartcard Technology 1999
Kellen Cheng, Anna Lisa Gentile, et al.
EMNLP 2024