TopInG: Topologically Interpretable Graph Learning via Persistent Rationale Filtration

Published in 42nd International Conference on Machine Learning (ICML), 2025

Abstract

Graph Neural Networks (GNNs) have shown remarkable success across various scientific fields, but their adoption in critical decision-making is often hindered by a lack of interpretability. Recently, intrinsically interpretable GNNs have been developed to provide insights into model predictions by identifying “rationale” substructures in graphs. However, existing methods struggle when these underlying rationale subgraphs are complex and varied in their structure.

To address this, we propose TopInG: Topologically Interpretable Graph Learning, a novel framework that leverages persistent homology to identify stable and meaningful rationale subgraphs. TopInG uses a rationale filtration learning approach to model the generation of these subgraphs. We introduce a self-adjusting topological constraint, which we term topological discrepancy, to enforce a clear and persistent topological distinction between the crucial rationale subgraphs and irrelevant parts of the graph. We provide theoretical guarantees that our loss function is uniquely optimized by the ground truth under specific conditions. Extensive experiments show that TopInG effectively handles variform rationales, balances predictive performance with interpretability, and mitigates spurious correlations, outperforming state-of-the-art methods in both prediction accuracy and interpretation quality.


The Challenge: Variform Rationales

Existing interpretable GNNs struggle with variform rationale subgraphs—explanatory substructures that vary significantly in shape, size, and topology within the same class. We created the BA-HouseOrGrid-nRnd dataset to test this challenge. As rationale complexity increases, existing methods degrade sharply while TopInG maintains stable performance.

Comparison on BA-HouseOrGrid-nRnd dataset

Figure 1: As rationale complexity (nRnd) increases, TopInG maintains high interpretation performance (ROC AUC) while other methods degrade significantly.


Our Method: Persistent Rationale Filtration

TopInG learns an optimal “filtration”—an ordered edge sequence based on importance. We first construct the rationale subgraph (GX), then add remaining edges (Gε). Persistent homology analyzes topological features that emerge during this process.

TopInG Method Overview

Figure 2: An overview of the TopInG framework.

Our key innovation is the topological discrepancy loss that maximizes differences between rationale and complement topological signatures, creating a clear “persistent homology gap.”

Persistent Homology Gap

Figure 3: Persistent homology barcode showing the topological discrepancy between rationale and complement subgraphs.


Experimental Results

We evaluated TopInG on eight benchmark datasets against leading interpretation methods, achieving new state-of-the-art results in both interpretation quality and predictive accuracy.

Interpretation Performance (AUC)

TopInG consistently identifies correct rationale subgraphs with higher fidelity. With the CINpp backbone, TopInG achieves perfect scores on challenging datasets.

Table 1: Interpretation Performance (AUC) on Test Datasets

MethodBA-2MotifsBA-HouseGridSPmotif0.5SPMotif0.9BA-HouseAndGridBA-HouseOrGridMutagBenzene
GNNExplainer67.35 ± 3.2950.73 ± 0.3462.62 ± 1.3558.85 ± 1.9353.04 ± 0.3853.21 ± 0.3661.98 ± 5.4548.72 ± 0.14
PGExplainer84.59 ± 9.0950.92 ± 1.5169.54 ± 5.6472.34 ± 2.9110.36 ± 4.373.14 ± 0.0160.91 ± 17.104.26 ± 0.36
MatchExplainer86.06 ± 28.3764.32 ± 2.3257.29 ± 14.3547.29 ± 13.3981.67 ± 0.4879.87 ± 1.6191.04 ± 6.5955.65 ± 1.16
Mage79.81 ± 2.2782.69 ± 4.7876.63 ± 0.9574.38 ± 0.6499.95 ± 0.0699.93 ± 0.0799.57 ± 0.4796.03 ± 0.63
Intrinsic Methods
DIR82.78 ± 10.9765.50 ± 15.3178.15 ± 1.3249.08 ± 3.6664.96 ± 14.3159.71 ± 21.5664.44 ± 28.8154.08 ± 13.75
GIN+GSAT98.85 ± 0.4798.55 ± 0.5974.49 ± 4.4665.25 ± 4.4292.92 ± 2.0383.56 ± 3.5799.38 ± 0.2591.57 ± 1.48
GIN+GMT-Lin97.72 ± 0.5985.68 ± 2.7976.26 ± 5.0769.08 ± 5.1476.12 ± 7.4774.36 ± 5.4199.87 ± 0.0983.90 ± 6.07
GIN+TopInG99.57* ± 0.6099.24* ± 0.6679.50 ± 3.7180.82* ± 4.2295.35 ± 0.9588.56 ± 2.0495.79 ± 1.9398.22* ± 0.92
CINpp Backbone
CINpp+GSAT91.12 ± 4.9391.04 ± 6.5978.20 ± 4.4880.24 ± 3.6695.17 ± 2.4669.30 ± 2.4897.27 ± 0.4795.40 ± 3.05
CINpp+GMT-Lin91.03 ± 5.2493.68 ± 4.7983.23 ± 4.3076.40 ± 2.3885.08 ± 3.8567.91 ± 5.1097.48 ± 0.8194.44 ± 2.49
CINpp+TopInG100.00* ± 0.0099.87* ± 0.1395.08* ± 1.8492.82* ± 2.45100.00* ± 0.00100.00* ± 0.0096.38 ± 2.56100.00* ± 0.00

*Results are statistically significant and outperform the best baseline.

Prediction Performance (Accuracy)

TopInG achieves higher prediction accuracy, especially with spurious correlations that mislead other models.

Table 2: Prediction Performance (Accuracy) on Test Datasets

ModelMethodMutagBenzeneSpuriousMotif (b=0.5)SpuriousMotif (b=0.7)SpuriousMotif (b=0.9)
GINDIR68.72 ± 2.5150.67 ± 0.9345.49 ± 3.8141.13 ± 2.6237.61 ± 2.02
GINGSAT98.28 ± 0.78100.00 ± 0.0047.45 ± 5.8743.57 ± 2.4345.39 ± 5.02
GINGMT-Lin91.20 ± 2.75100.00 ± 0.0051.16 ± 3.5153.11 ± 4.1247.60 ± 2.06
GINTopInG94.20 ± 5.61100.00 ± 0.0052.22 ± 2.0754.46 ± 5.7650.21 ± 3.22
CINpp Backbone
CINppGSAT96.14 ± 0.6799.43 ± 0.5474.70 ± 3.3770.41 ± 3.4465.90 ± 4.18
CINppGMT-Lin95.27 ± 1.3698.87 ± 0.9273.16 ± 3.5169.11 ± 4.1268.60 ± 6.06
CINppTopInG92.92 ± 7.02100.00* ± 0.0079.30* ± 3.9275.46* ± 4.6277.68* ± 4.64

*Results are statistically significant and outperform the best baseline.


Key Contributions

  • Novel Framework: First to leverage persistent homology for graph interpretability
  • Theoretical Guarantees: Topological discrepancy loss uniquely optimized by ground truth
  • Superior Performance: State-of-the-art interpretation quality and predictive accuracy
  • Robustness: Handles variform rationales and spurious correlations

Poster: Download PosterPDF: Download PDF