ARTFEED — Contemporary Art Intelligence

Self-Inconsistency in GNN Explanations and Self-Denoising Strategy

other · 2026-05-11

Researchers have discovered that the explanations generated by Self-Interpretable Graph Neural Networks (SI-GNNs) can exhibit self-inconsistency; when the model is applied again to its own explanatory subgraph, it produces varying explanations. This inconsistency is primarily due to context perturbation induced by re-explanation. A hypothesis regarding latent signal assignment clarifies why certain edges are more affected by this perturbation, while conciseness regularization plays a role in this assignment. The authors introduce Self-Denoising (SD), a training-free, model-agnostic post-processing approach that refines explanations with a single additional forward pass. Experiments conducted across various SI-GNN frameworks and backbone architectures validate its effectiveness.

Key facts

  • Self-inconsistency arises in SI-GNN explanations
  • Re-explanation-induced context perturbation is the direct cause
  • Latent signal assignment hypothesis explains edge sensitivity
  • Conciseness regularization affects latent signal assignment
  • Self-Denoising (SD) is a model-agnostic, training-free post-processing strategy
  • SD calibrates explanations with one additional forward pass
  • Experiments conducted across representative SI-GNN frameworks
  • Study appears on arXiv (2605.07527)

Entities

Institutions

  • arXiv

Sources