Evaluating the quality of machine learning explanations: A survey on methods and metrics

J Zhou, AH Gandomi, F Chen, A Holzinger - Electronics, 2021 - mdpi.com
The most successful Machine Learning (ML) systems remain complex black boxes to end-
users, and even experts are often unable to understand the rationale behind their decisions …

Towards a science of human-AI decision making: An overview of design space in empirical human-subject studies

V Lai, C Chen, A Smith-Renner, QV Liao… - Proceedings of the 2023 …, 2023 - dl.acm.org
AI systems are adopted in numerous domains due to their increasingly strong predictive
performance. However, in high-stakes domains such as criminal justice and healthcare, full …

Capabilities of gpt-4 on medical challenge problems

H Nori, N King, SM McKinney, D Carignan… - arXiv preprint arXiv …, 2023 - arxiv.org
Large language models (LLMs) have demonstrated remarkable capabilities in natural
language understanding and generation across various domains, including medicine. We …

[HTML][HTML] Explainable Artificial Intelligence (XAI): What we know and what is left to attain Trustworthy Artificial Intelligence

S Ali, T Abuhmed, S El-Sappagh, K Muhammad… - Information fusion, 2023 - Elsevier
Artificial intelligence (AI) is currently being utilized in a wide range of sophisticated
applications, but the outcomes of many AI models are challenging to comprehend and trust …

Explanations can reduce overreliance on ai systems during decision-making

H Vasconcelos, M Jörke… - Proceedings of the …, 2023 - dl.acm.org
Prior work has identified a resilient phenomenon that threatens the performance of human-
AI decision-making teams: overreliance, when people agree with an AI, even when it is …

“Everyone wants to do the model work, not the data work”: Data Cascades in High-Stakes AI

N Sambasivan, S Kapania, H Highfill… - proceedings of the …, 2021 - dl.acm.org
AI models are increasingly applied in high-stakes domains like health and conservation.
Data quality carries an elevated significance in high-stakes AI due to its heightened …

To trust or to think: cognitive forcing functions can reduce overreliance on AI in AI-assisted decision-making

Z Buçinca, MB Malaya, KZ Gajos - Proceedings of the ACM on Human …, 2021 - dl.acm.org
People supported by AI-powered decision support tools frequently overrely on the AI: they
accept an AI's suggestion even when that suggestion is wrong. Adding explanations to the …

Does the whole exceed its parts? the effect of ai explanations on complementary team performance

G Bansal, T Wu, J Zhou, R Fok, B Nushi… - Proceedings of the …, 2021 - dl.acm.org
Many researchers motivate explainable AI with studies showing that human-AI team
performance on decision-making tasks improves when the AI explains its recommendations …

Towards a science of human-ai decision making: a survey of empirical studies

V Lai, C Chen, QV Liao, A Smith-Renner… - arXiv preprint arXiv …, 2021 - arxiv.org
As AI systems demonstrate increasingly strong predictive performance, their adoption has
grown in numerous domains. However, in high-stakes domains such as criminal justice and …

" Help Me Help the AI": Understanding How Explainability Can Support Human-AI Interaction

SSY Kim, EA Watkins, O Russakovsky, R Fong… - Proceedings of the …, 2023 - dl.acm.org
Despite the proliferation of explainable AI (XAI) methods, little is understood about end-
users' explainability needs and behaviors around XAI explanations. To address this gap and …