On the Expressiveness and Length Generalization of Selective State-Space Models on Regular LanguagesAleksandar TerzicMichael Herscheet al.2025AAAI 2025
Efficient Scaling of Large Language Models with Mixture of Experts and 3D Analog In-Memory ComputingJulian BüchelA. Vasilopouloset al.2025Nat. Comput. Sci.
Limits of Transformer Language Models on Learning to Compose AlgorithmsJonathan ThommAleksandar Terzicet al.2024NeurIPS 2024
On the role of noise in factorizers for disentangling distributed representationsKumudu Geethan KarunaratneMichael Herscheet al.2024NeurIPS 2024
Recurrent Transformers Trade-off Parallelism for Length Generalization on Regular LanguagesPaul SoulosAleksandar Terzicet al.2024NeurIPS 2024
RETRO-LI: Small-Scale Retrieval Augmented Generation Supporting Noisy Similarity Searches and Domain Shift GeneralizationGentiana RashitiKumudu Geethan Karunaratneet al.2024ECAI 2024
Factorizers for Distributed Sparse Block CodesMichael HerscheAleksandar Terzicet al.2024Neurosymbolic AI
Towards Learning Abductive Reasoning using VSA Distributed RepresentationsGiacomo CamposampieroMichael Herscheet al.2024NeSy 2024
Exploiting Computation in Superposition via Multiple-Input-Multiple-Output Neural NetworksNicolas MenetMichael Herscheet al.2024NeSy 2024