Gewählte Publikation:
SHR
Neuro
Krebs
Kardio
Lipid
Stoffw
Microb
Müller, H; Holzinger, A; Plass, M; Brcic, L; Stumptner, C; Zatloukal, K.
Explainability and causability for artificial intelligence-supported medical image analysis in the context of the European In Vitro Diagnostic Regulation.
N Biotechnol. 2022; 70:67-72
Doi: 10.1016/j.nbt.2022.05.002
Web of Science
PubMed
FullText
FullText_MUG
- Führende Autor*innen der Med Uni Graz
-
Müller Heimo
-
Zatloukal Kurt
- Co-Autor*innen der Med Uni Graz
-
Brcic Luka
-
Holzinger Andreas
-
Plass Markus
-
Stumptner Cornelia
- Altmetrics:
- Dimensions Citations:
- Plum Analytics:
- Scite (citation analytics):
- Abstract:
- Artificial Intelligence (AI) for the biomedical domain is gaining significant interest and holds considerable potential for the future of healthcare, particularly also in the context of in vitro diagnostics. The European In Vitro Diagnostic Medical Device Regulation (IVDR) explicitly includes software in its requirements. This poses major challenges for In Vitro Diagnostic devices (IVDs) that involve Machine Learning (ML) algorithms for data analysis and decision support. This can increase the difficulty of applying some of the most successful ML and Deep Learning (DL) methods to the biomedical domain, just by missing the required explanatory components from the manufacturers. In this context, trustworthy AI has to empower biomedical professionals to take responsibility for their decision-making, which clearly raises the need for explainable AI methods. Explainable AI, such as layer-wise relevance propagation, can help in highlighting the relevant parts of inputs to, and representations in, a neural network that caused a result and visualize these relevant parts. In the same way that usability encompasses measurements for the quality of use, the concept of causability encompasses measurements for the quality of explanations produced by explainable AI methods. This paper describes both concepts and gives examples of how explainability and causability are essential in order to demonstrate scientific validity as well as analytical and clinical performance for future AI-based IVDs.
- Find related publications in this database (using NLM MeSH Indexing)
-
Algorithms - administration & dosage
-
Artificial Intelligence - administration & dosage
-
Machine Learning - administration & dosage
-
Neural Networks, Computer - administration & dosage
-
Software - administration & dosage
- Find related publications in this database (Keywords)
-
Medical AI
-
Retractability
-
Causability
-
Explainability
-
Scientific validity
-
Regulatory requirements
-
IVDR
-
In vitro diagnostic device regulation