Never tell me the odds: Investigating pro-hoc explanations in medical decision making

Federico Cabitza, Chiara Natali,Lorenzo Famiglini, Andrea Campagner, Valerio Caccavella,Enrico Gallazzi

ARTIFICIAL INTELLIGENCE IN MEDICINE(2024)

引用 0|浏览1
暂无评分
摘要
This paper examines a kind of explainable AI, centered around what we term pro -hoc explanations, that is a form of support that consists of offering alternative explanations (one for each possible outcome) instead of a specific post -hoc explanation following specific advice. Specifically, our support mechanism utilizes explanations by examples, featuring analogous cases for each category in a binary setting. Pro -hoc explanations are an instance of what we called frictional AI, a general class of decision support aimed at achieving a useful compromise between the increase of decision effectiveness and the mitigation of cognitive risks, such as over-reliance, automation bias and deskilling. To illustrate an instance of frictional AI, we conducted an empirical user study to investigate its impact on the task of radiological detection of vertebral fractures in x-rays. Our study engaged 16 orthopedists in a 'human -first, second -opinion' interaction protocol. In this protocol, clinicians first made initial assessments of the x-rays without AI assistance and then provided their final diagnosis after considering the pro -hoc explanations. Our findings indicate that physicians, particularly those with less experience, perceived pro -hoc XAI support as significantly beneficial, even though it did not notably enhance their diagnostic accuracy. However, their increased confidence in final diagnoses suggests a positive overall impact. Given the promisingly high effect size observed, our results advocate for further research into pro -hoc explanations specifically, and into the broader concept of frictional AI.
更多
查看译文
关键词
eXplainable AI,Decision support,Machine learning,Frictional AI
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要