Publication
DATE 2024
Invited talk

Reducing computational complexity of perception and reasoning by neuro-vector-symbolic architectures

Abstract

We recently proposed neuro-vector-symbolic architectures (NVSA) in which high-dimensional distributed vectors are properly generated by neural nets and further processed by a VSA-informed machinery at different levels of abstraction. Using NVSA, we could set state-of-the-art accuracy record on few-shot continual learning [CVPR 2022] as well as visual abstract reasoning tasks [Nature Machine Intelligence 2023]. This is not where the advantages end: NVSA also reduces the computational complexity associated with both perceptual and reasoning tasks, yet on modern CPUs/GPUs. NVSA expanded computation-in-superposition to highly nonlinear transformations in CNNs and Transformers, effectively doubling their throughput at nearly iso-accuracy and computational cost [NeurIPS 2023]. NVSA also made probabilistic abduction tractable by avoiding exhaustive probability computations and brute-force symbolic searches which led to 244× faster inference compared with the probabilistic reasoning within the state-of-the-art approaches [Nature Machine Intelligence 2023]. Finally, NVSA permitted learning-to-reason: instead of hard-coding the rule formulations associated with a reasoning task, NVSA could transparently learn the rule formulations with just one pass through the training data [NeurIPSW Math-AI 2023].

Date

25 Mar 2024

Publication

DATE 2024