학술논문

Explicability of artificial intelligence in radiology: Is a fifth bioethical principle conceptually necessary?
Document Type
Report
Source
Bioethics. February, 2022, Vol. 36 Issue 2, p143, 11 p.
Subject
Artificial intelligence -- Analysis
Radiology -- Analysis
Machine learning -- Analysis
Medical ethics -- Analysis
Data mining -- Analysis
Radiology, Medical -- Analysis
Data warehousing/data mining
Artificial intelligence
Biological sciences
Philosophy and religion
Language
English
ISSN
0269-9702
Abstract
Keywords: black box; explainability; machine learning; medical ethics; principlism; transparency Abstract Recent years have witnessed intensive efforts to specify which requirements ethical artificial intelligence (AI) must meet. General guidelines for ethical AI consider a varying number of principles important. A frequent novel element in these guidelines, that we have bundled together under the term explicability, aims to reduce the black-box character of machine learning algorithms. The centrality of this element invites reflection on the conceptual relation between explicability and the four bioethical principles. This is important because the application of general ethical frameworks to clinical decision-making entails conceptual questions: Is explicability a free-standing principle? Is it already covered by the well-established four bioethical principles? Or is it an independent value that needs to be recognized as such in medical practice? We discuss these questions in a conceptual-ethical analysis, which builds upon the findings of an empirical document analysis. On the example of the medical specialty of radiology, we analyze the position of radiological associations on the ethical use of medical AI. We address three questions: Are there references to explicability or a similar concept? What are the reasons for such inclusion? Which ethical concepts are referred to? Byline: Nils-Frederic Wagner, Mita Banerjee, Norbert W. Paul, Frank Ursin, Cristian Timmermann, Florian Steger