Cost-Aware Counterfactuals for Black Box Explanations
Natalia Martinez Gil, Kanthi Sarpatwar, et al.
NeurIPS 2023
In time-series classification, interpretable models can bring additional insights but be outperformed by deep models since human-understandable features have limited expressivity and flexibility. In this work, we present InterpGN, a framework that integrates an interpretable model and a deep neural network. Within this framework, we introduce a novel gating function design based on the confidence of the interpretable expert, preserving interpretability for samples where interpretable features are significant while also identifying samples that require additional expertise. For the interpretable expert, we incorporate shapelets to effectively model shape-level features for time-series data. We introduce a variant of Shapelet Transforms to build logical predicates using shapelets. Our proposed model achieves comparable performance with state-of-the-art deep learning models while additionally providing interpretable classifiers for various benchmark datasets. We further show that our models improve on quantitative shapelet quality and interpretability metrics over existing shapelet-learning formulations. Finally, we show that our models can integrate additional advanced architectures and be applied to real-world tasks beyond standard benchmarks such as the MIMIC-III and time series extrinsic regression datasets.
Natalia Martinez Gil, Kanthi Sarpatwar, et al.
NeurIPS 2023
Gang Liu, Michael Sun, et al.
ICLR 2025
Daniel Karl I. Weidele, Hendrik Strobelt, et al.
SysML 2019
Michael Hersche, Francesco Di Stefano, et al.
NeurIPS 2023