Interpretable deep learning: Interpretation, interpretability, trustworthiness, and beyond

X Li, H Xiong, X Li, X Wu, X Zhang, J Liu, J Bian… - … and Information Systems, 2022 - Springer
Deep neural networks have been well-known for their superb handling of various machine
learning and artificial intelligence tasks. However, due to their over-parameterized black-box …

Counterfactual explanations and algorithmic recourses for machine learning: A review

S Verma, V Boonsanong, M Hoang, K Hines… - ACM Computing …, 2024 - dl.acm.org
Machine learning plays a role in many deployed decision systems, often in ways that are
difficult or impossible to understand by human stakeholders. Explaining, in a human …

Deep neural networks and tabular data: A survey

V Borisov, T Leemann, K Seßler, J Haug… - IEEE transactions on …, 2022 - ieeexplore.ieee.org
Heterogeneous tabular data are the most commonly used form of data and are essential for
numerous critical and computationally demanding applications. On homogeneous datasets …

A survey of uncertainty in deep neural networks

J Gawlikowski, CRN Tassi, M Ali, J Lee, M Humt… - Artificial Intelligence …, 2023 - Springer
Over the last decade, neural networks have reached almost every field of science and
become a crucial part of various real world applications. Due to the increasing spread …

Interpretable machine learning: Fundamental principles and 10 grand challenges

C Rudin, C Chen, Z Chen, H Huang… - Statistic …, 2022 - projecteuclid.org
Interpretability in machine learning (ML) is crucial for high stakes decisions and
troubleshooting. In this work, we provide fundamental principles for interpretable ML, and …

Uncertainty as a form of transparency: Measuring, communicating, and using uncertainty

U Bhatt, J Antorán, Y Zhang, QV Liao… - Proceedings of the …, 2021 - dl.acm.org
Algorithmic transparency entails exposing system properties to various stakeholders for
purposes that include understanding, improving, and contesting predictions. Until now, most …

Explaining in style: Training a gan to explain a classifier in stylespace

O Lang, Y Gandelsman, M Yarom… - Proceedings of the …, 2021 - openaccess.thecvf.com
Image classification models can depend on multiple different semantic attributes of the
image. An explanation of the decision of the classifier needs to both discover and visualize …

Uncertainty quantification with pre-trained language models: A large-scale empirical analysis

Y Xiao, PP Liang, U Bhatt, W Neiswanger… - arXiv preprint arXiv …, 2022 - arxiv.org
Pre-trained language models (PLMs) have gained increasing popularity due to their
compelling prediction performance in diverse natural language processing (NLP) tasks …

Carla: a python library to benchmark algorithmic recourse and counterfactual explanation algorithms

M Pawelczyk, S Bielawski, J Heuvel, T Richter… - arXiv preprint arXiv …, 2021 - arxiv.org
Counterfactual explanations provide means for prescriptive model explanations by
suggesting actionable feature changes (eg, increase income) that allow individuals to …

Sample-efficient optimization in the latent space of deep generative models via weighted retraining

A Tripp, E Daxberger… - Advances in Neural …, 2020 - proceedings.neurips.cc
Many important problems in science and engineering, such as drug design, involve
optimizing an expensive black-box objective function over a complex, high-dimensional, and …