
A Comprehensive Explanation Framework for Biomedical Time Series Classification
Author(s) -
Praharsh Ivaturi,
Matteo Gadaleta,
Amitabh C. Pandey,
Michael Pazzani,
Steven R. Steinhubl,
Giorgio Quer
Publication year - 2021
Publication title -
ieee journal of biomedical and health informatics
Language(s) - English
Resource type - Journals
SCImago Journal Rank - 1.293
H-Index - 125
eISSN - 2168-2208
pISSN - 2168-2194
DOI - 10.1109/jbhi.2021.3060997
Subject(s) - bioengineering , communication, networking and broadcast technologies , components, circuits, devices and systems , computing and processing , signal processing and analysis
In this study, we propose a post-hoc explainability framework for deep learning models applied to quasi-periodic biomedical time-series classification. As a case study, we focus on the problem of atrial fibrillation (AF) detection from electrocardiography signals, which has strong clinical relevance. Starting from a state-of-the-art pretrained model, we tackle the problem from two different perspectives: global and local explanation. With global explanation, we analyze the model behavior by looking at entire classes of data, showing which regions of the input repetitive patterns have the most influence for a specific outcome of the model. Our explanation results align with the expectations of clinical experts, showing that features crucial for AF detection contribute heavily to the final decision. These features include R-R interval regularity, absence of the P-wave or presence of electrical activity in the isoelectric period. On the other hand, with local explanation, we analyze specific input signals and model outcomes. We present a comprehensive analysis of the network facing different conditions, whether the model has correctly classified the input signal or not. This enables a deeper understanding of the network's behavior, showing the most informative regions that trigger the classification decision and highlighting possible causes of misbehavior.