Abstract
Graph Neural Networks (GNNs) are a popular approach for predicting graph structured data. As GNNs tightly entangle the input graph into the neural network structure, common explainable AI approaches are not applicable. To a large extent, GNNs have remained black-boxes for the user so far. In this paper, we show that GNNs can in fact be naturally explained using higher-order expansions, i.e., by identifying groups of edges that jointly contribute to the prediction. Practically, we find that such explanations can be extracted using a nested attribution scheme, where existing techniques such as layer-wise relevance propagation (LRP) can be applied at each step. The output is a collection of walks into the input graph that are relevant for the prediction. Our novel explanation method, which we denote by GNN-LRP, is applicable to a broad range of graph neural networks and lets us extract practically relevant insights on sentiment analysis of text data, structure-property relationships in quantum chemistry, and image classification.
| Original language | English |
|---|---|
| Pages (from-to) | 7581-7596 |
| Number of pages | 16 |
| Journal | IEEE Transactions on Pattern Analysis and Machine Intelligence |
| Volume | 44 |
| Issue number | 11 |
| DOIs | |
| Publication status | Published - 2022 Nov 1 |
Bibliographical note
Publisher Copyright:© 1979-2012 IEEE.
Keywords
- explainable machine learning
- Graph neural networks
- higher-order explanations
- layer-wise relevance propagation
ASJC Scopus subject areas
- Software
- Computer Vision and Pattern Recognition
- Computational Theory and Mathematics
- Artificial Intelligence
- Applied Mathematics
Fingerprint
Dive into the research topics of 'Higher-Order Explanations of Graph Neural Networks via Relevant Walks'. Together they form a unique fingerprint.Cite this
- APA
- Standard
- Harvard
- Vancouver
- Author
- BIBTEX
- RIS