Graph Reconstruction from Differentially Private GNN Explanations
Rishi Raj Sahoo, Jyotirmaya Shivottam, Subhankar Mishra
TLDR
This paper reveals that differentially private GNN explanations can still leak significant graph structure, proposing PRIVX, a diffusion-based attack.
Key contributions
- Presents PRIVX, a diffusion-based attack to reconstruct graph structure from differentially private GNN explanations.
- Shows significant graph structure leakage from DP-perturbed GNN explanations, achieving high AUC under typical privacy budgets.
- Offers guidance on explainer choice, detailing how leakage varies with graph homophily/heterophily and explainer type.
Why it matters
This paper highlights a critical privacy vulnerability in releasing differentially private GNN explanations, showing that graph structure can still be reconstructed. It provides a novel attack method and practical recommendations for practitioners to mitigate this unexpected data leakage.
Original Abstract
Regulatory frameworks such as GDPR increasingly require that ML predictions be accompanied by post-hoc explanations, even when raw data and trained models cannot be released. Differential privacy (DP) is the standard mitigation for the residual privacy risk of releasing these explanations. We show that DP is not sufficient: an adversary observing only DP-perturbed GNN explanations can reconstruct hidden graph structure with high accuracy. Our attack, PRIVX, exploits the fact that the Gaussian DP mechanism is a single DDPM forward step at known noise level σ(ε), recasting reconstruction as reverse diffusion conditioned on the corrupted signal, a principled Bayesian denoiser under known DP corruption. We formalise a stratified adversary model parameterised by (M, \hatε, \hatδ, S, ρ) that interpolates between oblivious and oracle attackers, and derive endpoint-matched two-sided bounds on reconstruction AUC. For practitioners, we provide regime-stratified guidance on explainer choice: on homophilic graphs, neighbourhood-aggregating explainers (GraphLIME, GNNExplainer) leak more structure than per-node gradient explainers under the same DP budget; on strongly heterophilic graphs the ordering reverses. We introduce PRIVF as an auxiliary diagnostic sharing the same diffusion backbone to decompose leakage into explainer-induced and intrinsic graph-distribution components. Experiments across seven benchmarks, three DP mechanisms, and three GNN backbones show PRIVX achieves AUC above 0.7 at ε = 5 on five of seven datasets, with the attack succeeding well within typically deployed privacy budgets.
📬 Weekly AI Paper Digest
Get the top 10 AI/ML arXiv papers from the week — summarized, scored, and delivered to your inbox every Monday.