“…Interpretability of image-based classification models is absolutely necessary in biomedical domains where mechanistic understanding and transparency are crucial. Established attribution-based (Barnett et al 2021, Kraus et al 2017, Graziani et al 2018, Wu et al 2018, Singh et al 2020, Zhang et al 2021) or counterfactual-explanation based ( Singla et al 2023 , Thiagarajan et al 2022, Mertes et al 2022, Narayanaswamy et al 2020, Soelistyo et al 2022, Zaritsky et al 2021, Lamiable et al 2023, Kraus et al 2017) methods were applied, out-of the box or after some adaptations, to interpret a variety of biomedical image-based classification tasks. DISCOVER’s classification-driven and disentanglement representations overcome the inherent limitations in these methods and enabled us to quantitatively confirm non-trivial interpretations, rather than relying on qualitative explanations of representative images, and to systematically perform quantitative instance-specific interpretations.…”