Training-free Counterfactual Explanation for Temporal Graph Model Inference

ICLR 2026 Conference SubmissionAnonymous Authors
Temporal Graph Nerual NetworksExplainabilityTraining Free
Abstract:

Temporal graph neural networks (TGNN) extend graph neural networks to dynamic networks and have demonstrated strong predictive power. However, interpreting TGNN remains far less explored than their static-graph counterparts. This paper introduces TEMporal Graph eXplainer (TemGX), a training-free,post-hoc framework that help users interpret and understand TGNN behavior by discovering temporal subgraphs and their evolution that are responsible for TGNN output of interests.We introduce a class of explainability measures that extends influence maximization in terms of structural influence and time decay to model temporal influence. We formulate the explanation task as a constrained optimization problem, and propose fast algorithms to discover explanations with guarantees on their temporal explainability. Our experimental study verifies the effectiveness and efficiency of TemGX for TGNN explanation, compared with state-of-the-art explainers. We also showcase how TemGX supports inference queries for dynamic network analysis.

Disclaimer
This report is AI-GENERATED using Large Language Models and WisPaper (A scholar search engine). It analyzes academic papers' tasks and contributions against retrieved prior work. While this system identifies POTENTIAL overlaps and novel directions, ITS COVERAGE IS NOT EXHAUSTIVE AND JUDGMENTS ARE APPROXIMATE. These results are intended to assist human reviewers and SHOULD NOT be relied upon as a definitive verdict on novelty.
NOTE that some papers exist in multiple, slightly different versions (e.g., with different titles or URLs). The system may retrieve several versions of the same underlying work. The current automated pipeline does not reliably align or distinguish these cases, so human reviewers will need to disambiguate them manually.
If you have any questions, please contact: mingzhang23@m.fudan.edu.cn

Overview

Taxonomy

Core-task Taxonomy Papers
50
3
Claimed Contributions
21
Contribution Candidate Papers Compared
0
Refutable Paper

Research Landscape Overview

Core task: Explaining temporal graph neural network predictions. The field of explainability for temporal graph neural networks has evolved into several distinct branches that reflect different methodological philosophies and application contexts. Post-hoc explanation methods form a major branch, encompassing perturbation-based and influence-based explainers that analyze trained models after the fact, including counterfactual and causal explanation techniques such as DyExplainer[2] and Causality-inspired Explanations[3]. Self-interpretable architectures represent an alternative paradigm, embedding transparency directly into model design through mechanisms like attention or bottleneck structures, as seen in Self-explainable Bottleneck[14]. Domain-specific applications demonstrate how these techniques are tailored to particular fields such as healthcare (Interpretable Alzheimer Analysis[4], ADHD Biomarkers[6]), urban systems (Traffexplainer[33]), and infrastructure monitoring. Theoretical foundations and frameworks provide the mathematical underpinnings, including work on Koopman Theory[7] and evaluation methodologies like Evaluating Explainability[25]. Specialized temporal graph architectures focus on novel representational approaches that facilitate both prediction and interpretation, such as continuous-time models and higher-order structures. Within the post-hoc methods branch, a particularly active line of work explores counterfactual and causal reasoning to identify which temporal graph features drive predictions. Training-free Counterfactual[0] sits squarely in this cluster, emphasizing efficient counterfactual generation without requiring model retraining. This contrasts with nearby approaches like Causality-inspired Explanations[3], which may incorporate more explicit causal modeling frameworks, and Dynamic Causal[15], which focuses on evolving causal structures over time. A key tension across these methods involves balancing computational efficiency with the depth of causal insight: some works prioritize scalable perturbation strategies, while others invest in richer causal representations. Training-free Counterfactual[0] appears to lean toward the efficiency side, offering a lightweight alternative within a landscape where methods like DyExplainer[2] and its sparse variant DyExplainer Sparse[8] have established benchmarks for perturbation-based temporal explanations.

Claimed Contributions

TemGX framework for training-free TGNN explanation

The authors propose TemGX, a framework that discovers temporal subgraphs and their evolution responsible for TGNN inference results through counterfactual analysis. The framework is training-free, model-agnostic, and supports queries for interpretable dynamic network analysis.

4 retrieved papers
Temporal explainability measures integrating spatial-temporal influence and time decay

The authors develop explainability measures that combine spatial-temporal influence propagation with time decay modeling to quantify temporal influence within sliding windows. These measures extend information cascading models and resistance distance analysis for evolving graphs.

7 retrieved papers
Efficient algorithms with approximation guarantees for temporal explanation discovery

The authors formulate TGNN explanation as a constrained optimization problem and develop efficient algorithms that discover temporal explanations with provable approximation guarantees. The algorithm achieves a (1 - 1/e)-approximation ratio for the temporal explainability measure.

10 retrieved papers

Core Task Comparisons

Comparisons with papers in the same taxonomy category

Contribution Analysis

Detailed comparisons for each claimed contribution

Contribution

TemGX framework for training-free TGNN explanation

The authors propose TemGX, a framework that discovers temporal subgraphs and their evolution responsible for TGNN inference results through counterfactual analysis. The framework is training-free, model-agnostic, and supports queries for interpretable dynamic network analysis.

Contribution

Temporal explainability measures integrating spatial-temporal influence and time decay

The authors develop explainability measures that combine spatial-temporal influence propagation with time decay modeling to quantify temporal influence within sliding windows. These measures extend information cascading models and resistance distance analysis for evolving graphs.

Contribution

Efficient algorithms with approximation guarantees for temporal explanation discovery

The authors formulate TGNN explanation as a constrained optimization problem and develop efficient algorithms that discover temporal explanations with provable approximation guarantees. The algorithm achieves a (1 - 1/e)-approximation ratio for the temporal explainability measure.