Gewählte Publikation:
SHR
Neuro
Krebs
Kardio
Lipid
Stoffw
Microb
Finzel, B; Saranti, A; Angerschmid, A; Tafler, D; Pfeifer, B; Holzinger, A.
Generating Explanations for Conceptual Validation of Graph Neural Networks
KUNSTL INTELL. 2022;
Doi: 10.1007/s13218-022-00781-7
[OPEN ACCESS]
Web of Science
PubMed
FullText
FullText_MUG
- Führende Autor*innen der Med Uni Graz
-
Holzinger Andreas
- Co-Autor*innen der Med Uni Graz
-
Angerschmid Alessa
-
Pfeifer Bastian
-
Saranti Anna
- Altmetrics:
- Dimensions Citations:
- Plum Analytics:
- Scite (citation analytics):
- Abstract:
- Graph Neural Networks (GNN) show good performance in relational data classification. However, their contribution to concept learning and the validation of their output from an application domain's and user's perspective have not been thoroughly studied. We argue that combining symbolic learning methods, such as Inductive Logic Programming (ILP), with statistical machine learning methods, especially GNNs, is an essential forward-looking step to perform powerful and validatable relational concept learning. In this contribution, we introduce a benchmark for the conceptual validation of GNN classification outputs. It consists of the symbolic representations of symmetric and non-symmetric figures that are taken from a well-known Kandinsky Pattern data set. We further provide a novel validation framework that can be used to generate comprehensible explanations with ILP on top of the relevance output of GNN explainers and human-expected relevance for concepts learned by GNNs. Our experiments conducted on our benchmark data set demonstrate that it is possible to extract symbolic concepts from the most relevant explanations that are representative of what a GNN has learned. Our findings open up a variety of avenues for future research on validatable explanations for GNNs.
- Find related publications in this database (Keywords)
-
Graph neural networks (GNN)
-
Explainable AI (xAI)
-
Inductive logic programming (ILP)
-
Symbolic AI
-
Kandinsky pattern (KP)