Analog AI Accelerators for Transformer-based Language Models: Hardware, Workload, and Power PerformanceSidney TsaiHadjer Benmezianeet al.2025IMW 2025
Analog-AI Hardware Accelerators for low-latency Transformer-based Language Models (Invited)Geoffrey BurrSidney Tsaiet al.2025CICC 2025
NORA: Noise-Optimized Rescaling of LLMs on Analog Compute-in-Memory AcceleratorsYayue HouSidney Tsaiet al.2025DATE 2025
Multi-Task Neural Network Mapping onto Analog-Digital Heterogeneous AcceleratorsHadjer BenmezianeCorey Liam Lammieet al.2024NeurIPS 2024
AIHWKIT-Lightning: A Scalable HW-Aware Training Toolkit for Analog In-Memory ComputingJulian BüchelWilliam Simonet al.2024NeurIPS 2024
Medical Neural Architecture Search: Survey and TaxonomyHadjer BenmezianeImane Hamzaouiet al.2024IJCAI 2024
Grassroots operator search for model edge adaptation using mathematical search spaceHadjer BenmezianeKaoutar El Maghraouiet al.2024Future Generation Computer Systems
A Provably Effective Method for Pruning Experts in Fine-tuned Sparse Mixture-of-ExpertsMohammed Nowaz Rabbani ChowdhuryMeng Wanget al.2024ICML 2024
EfficientMedSAM: Accelerating Medical Image Segmentation via Neural Architecture Search and Knowledge DistillationArnold AsiimweWilliam Daset al.2024EDGE 2024
Are Large Language Models Good Neural Architecture Generators for Edge?Hadjer BenmezianeKaoutar El Maghraoui2024EDGE 2024