Analog AI Accelerators for Transformer-based Language Models: Hardware, Workload, and Power PerformanceSidney TsaiHadjer Benmezianeet al.2025IMW 2025
NORA: Noise-Optimized Rescaling of LLMs on Analog Compute-in-Memory AcceleratorsYayue HouSidney Tsaiet al.2025DATE 2025
Multi-Task Neural Network Mapping onto Analog-Digital Heterogeneous AcceleratorsHadjer BenmezianeCorey Liam Lammieet al.2024NeurIPS 2024
AIHWKIT-Lightning: A Scalable HW-Aware Training Toolkit for Analog In-Memory ComputingJulian BüchelWilliam Simonet al.2024NeurIPS 2024
Medical Neural Architecture Search: Survey and TaxonomyHadjer BenmezianeImane Hamzaouiet al.2024IJCAI 2024
Grassroots operator search for model edge adaptation using mathematical search spaceHadjer BenmezianeKaoutar El Maghraouiet al.2024Future Generation Computer Systems
A Provably Effective Method for Pruning Experts in Fine-tuned Sparse Mixture-of-ExpertsMohammed Nowaz Rabbani ChowdhuryMeng Wanget al.2024ICML 2024
EfficientMedSAM: Accelerating Medical Image Segmentation via Neural Architecture Search and Knowledge DistillationArnold AsiimweWilliam Daset al.2024EDGE 2024
Are Large Language Models Good Neural Architecture Generators for Edge?Hadjer BenmezianeKaoutar El Maghraoui2024EDGE 2024
Analog AI as a Service: A Cloud Platform for In-Memory ComputingKaoutar El MaghraouiKim Tranet al.2024SSE 2024