Self-Consistency Improves the Trustworthiness of Self-Interpretable GNNs
Overview
Overall Novelty Assessment
The paper proposes a self-consistency training strategy to improve explanation quality in self-interpretable GNNs by directly optimizing faithfulness during training. It resides in the 'Faithfulness-Driven Training Strategies' leaf, which contains only three papers total including this work. This represents a relatively sparse research direction within the broader taxonomy of 50 papers across 36 topics, suggesting the specific focus on training-time faithfulness optimization remains underexplored compared to architectural innovations or post-hoc evaluation methods that dominate other branches.
The taxonomy reveals neighboring work in 'Causal and Information-Theoretic Training Objectives' and 'Pre-Training and Transfer Learning for Interpretability', indicating alternative approaches to improving explanation quality through different training paradigms. The sibling papers in the same leaf address faithfulness through sufficient-necessary explanation decomposition and semantic-level supervision, representing distinct technical strategies within the shared goal of aligning training objectives with explanation criteria. The broader 'Training Objectives and Optimization' branch remains less populated than 'Self-Interpretable GNN Architectures' or 'Post-Hoc Explanation Methods', highlighting a gap between architectural design and training methodology research.
Among 24 candidates examined across three contributions, none were identified as clearly refuting the proposed work. The first contribution linking faithfulness to self-consistency examined 10 candidates with no refutations, the self-consistency training strategy examined 4 candidates with no refutations, and the empirical analysis of inconsistency patterns examined 10 candidates with no refutations. This suggests that within the limited search scope, the specific framing of faithfulness as self-consistency and the proposed training approach appear relatively novel, though the analysis does not claim exhaustive coverage of all potentially relevant prior work.
Based on the limited literature search of 24 semantically similar papers, the work appears to occupy a distinct position by connecting faithfulness optimization to self-consistency principles during training. The sparse population of the faithfulness-driven training leaf and absence of identified overlapping work within the examined candidates suggest potential novelty, though the restricted search scope means additional relevant work may exist beyond the top-K semantic matches and citation expansion performed.
Taxonomy
Research Landscape Overview
Claimed Contributions
The authors establish a theoretical connection showing that faithfulness in GNN explanations is fundamentally related to self-consistency. This insight enables faithfulness to be directly optimized during training rather than only evaluated post-hoc.
The authors propose a two-step training framework that adds a self-consistency loss to enforce agreement between successive explanations. This strategy is model-agnostic, requires no architectural changes, and can be applied to existing SI-GNNs.
The authors conduct empirical studies demonstrating that self-inconsistency in SI-GNN explanations primarily arises from unimportant features. This finding connects self-inconsistency to the redundancy problem identified in prior work and motivates their training approach.
Core Task Comparisons
Comparisons with papers in the same taxonomy category
[22] Self-interpretable graph learning with sufficient and necessary explanations PDF
[23] SES: Bridging the gap between explainability and prediction of graph neural networks PDF
Contribution Analysis
Detailed comparisons for each claimed contribution
Linking faithfulness to explanation self-consistency
The authors establish a theoretical connection showing that faithfulness in GNN explanations is fundamentally related to self-consistency. This insight enables faithfulness to be directly optimized during training rather than only evaluated post-hoc.
[6] Towards self-interpretable graph-level anomaly detection PDF
[43] Semantic Interpretation and Validation of Graph Attention-Based Explanations for GNN Models PDF
[55] Towards Faithful and Consistent Explanations for Graph Neural Networks PDF
[61] On Consistency in Graph Neural Network Interpretation PDF
[64] AttenhERG: a reliable and interpretable graph neural network framework for predicting hERG channel blockers PDF
[65] The intelligible and effective graph neural additive network PDF
[66] Reliable interpretability of biology-inspired deep neural networks PDF
[67] GraphXAIN: Narratives to Explain Graph Neural Networks PDF
[68] Cooperative Explanations of Graph Neural Networks PDF
[69] Chemistry-intuitive explanation of graph neural networks for molecular property prediction with substructure masking PDF
Self-consistency training strategy for SI-GNNs
The authors propose a two-step training framework that adds a self-consistency loss to enforce agreement between successive explanations. This strategy is model-agnostic, requires no architectural changes, and can be applied to existing SI-GNNs.
[55] Towards Faithful and Consistent Explanations for Graph Neural Networks PDF
[61] On Consistency in Graph Neural Network Interpretation PDF
[62] Multi-task multi-station earthquake monitoring: An all-in-one seismic Phase picking, Location, and Association Network (PLAN) PDF
[63] Interpretable graph neural network framework for ultra-low-power junctionless GAA FET current mirrors: bridging physics-based modeling and circuit design PDF
Empirical analysis of self-inconsistency patterns
The authors conduct empirical studies demonstrating that self-inconsistency in SI-GNN explanations primarily arises from unimportant features. This finding connects self-inconsistency to the redundancy problem identified in prior work and motivates their training approach.