Buckets:
Title: A Cognac Shot To Forget Bad Memories: Corrective Unlearning for Graph Neural Networks
URL Source: https://arxiv.org/html/2412.00789
Markdown Content: Back to arXiv
This is experimental HTML to improve accessibility. We invite you to report rendering errors. Use Alt+Y to toggle on accessible reporting links and Alt+Shift+Y to toggle off. Learn more about this project and help improve conversions.
Why HTML? Report Issue Back to Abstract Download PDF Abstract 1Introduction 2Corrective Unlearning for Graph Neural Networks 3Our Method: Cognac 4Experimental Setup 5Results & Discussion 6Related Work 7Limitations and Conclusion References
HTML conversions sometimes display errors due to content that did not convert correctly from the source. This paper uses the following packages that are not yet supported by the HTML conversion tool. Feedback on these issues are not necessary; they are known and are being worked on.
failed: fontawesome5 failed: etoc
Authors: achieve the best HTML results from your LaTeX submissions by following these best practices.
License: CC BY 4.0 arXiv:2412.00789v4 [cs.LG] 09 Jun 2025 A Cognac Shot To Forget Bad Memories: Corrective Unlearning for Graph Neural Networks Varshita Kolipaka Akshit Sinha Debangan Mishra Sumit Kumar Arvindh Arun Shashwat Goel Ponnurangam Kumaraguru Abstract
Graph Neural Networks (GNNs) are increasingly being used for a variety of ML applications on graph data. Because graph data does not follow the independently and identically distributed (i.i.d.) assumption, adversarial manipulations or incorrect data can propagate to other data points through message passing, which deteriorates the model’s performance. To allow model developers to remove the adverse effects of manipulated entities from a trained GNN, we study the recently formulated problem of Corrective Unlearning. We find that current graph unlearning methods fail to unlearn the effect of manipulations even when the whole manipulated set is known. We introduce a new graph unlearning method, Cognac, which can unlearn the effect of the manipulation set even when only 5% of it is identified. It recovers most of the performance of a strong oracle with fully corrected training data, even beating retraining from scratch without the deletion set, and is 8 x more efficient while also scaling to large datasets. We hope our work assists GNN developers in mitigating harmful effects caused by issues in real-world data, post-training.
ICML, Cognac, Corrective Unlearning, Machine Unlearning, Graphs, GNN, Graph Unlearning
\faGlobe cognac-gnn-unlearning.github.io \faGithub corrective-unlearning-for-gnns
1Introduction
Graph Neural Networks (GNNs) are seeing widespread adoption across diverse domains, from recommender systems to drug discovery (Wu et al., 2022; Zhang et al., 2022). Recently, GNNs have been scaled to large training sets for various graph foundation models (Mao et al., 2024; Arun et al., 2025). However, in these large-scale settings, it is prohibitively expensive to verify the integrity of every sample in the training data that can potentially affect desiderata like fairness (Konstantinov & Lampert, 2022), robustness (Paleka & Sanyal, 2023; Günnemann, 2022), and accuracy (Sanyal et al., 2021).
Making the training process robust to minority populations (Günnemann, 2022; Jin et al., 2020) is challenging and can adversely affect fairness and accuracy (Sanyal et al., 2022). Consequently, model developers may want post-hoc ways to remove the adverse impact of manipulated training data if they observe problematic model behavior on specific distributions of test-time inputs. Such an approach follows the recent trend of using post-training interventions to ensure models behave in intended ways (Ouyang et al., 2022). Recently, Goel et al. (2024) formulated corrective unlearning as the challenge of removing adverse effects of manipulated data with access to only a representative subset for unlearning while being agnostic to the type of manipulations. We study this problem in the context of GNNs, which face unique challenges due to the graph structure. The traditional assumption of independent and identically distributed (i.i.d.) samples does not hold for GNNs, as they use a message-passing mechanism that aggregates information from neighbors. This process makes GNNs vulnerable to adversarial perturbations, where modifying even a few nodes can propagate changes across large portions of the graph and result in widespread changes in model predictions (Bojchevski & Günnemann, 2019b; Zügner et al., 2018). Consequently, for GNNs to effectively unlearn, they must remove the influence of manipulated entities on their neighbors.
Figure 1:Illustration of our method Cognac. Initially (Left), the model is trained on manipulated data (Devils), out of which only a subset is identified for deletion (Dark-red-devils). Our method alternates between two steps. (1) Identifying neighbors by the deletion set, which can include both nodes from the remaining data (light red) and unidentified manipulated nodes (Purple), and pushes their representation away from the deletion set and toward other nodes in the neighborhood. (2) We then perform ascent on the deletion set labels and descent on the remaining data with separate optimizer instances. This cleanly separates the embeddings of the affected classes (Right), recovering the accuracy on the affected distribution, and maintaining it on the remaining distribution.
Corrective Unlearning is an emerging paradigm that focuses on removing the influence of arbitrary training data manipulations on a trained model, using only a representative subset of the manipulated data (Goel et al., 2024). In this work, we focus on the use of GNNs in node classification tasks, studying unlearning for targeted binary class confusion attacks (Lingam et al., 2024) on both edges and nodes. For edge unlearning, we evaluate the unlearning of spurious edges that change the graph topology in a way that violates the homophily assumption that most GNNs rely on. For node unlearning, we utilize a label flip attack (Lingam et al., 2024) which is used as a classical graph adversarial attack, similar to the Interclass Confusion attack (Goel et al., 2022).
First, we evaluate whether existing GNN unlearning methods are effective in removing the impact of manipulated entities. Our findings reveal that these methods consistently fail, even when provided with a complete set of manipulated entities. We then propose our method, Cognac, which unlearns by alternating between two components, as illustrated in Figure 1. The first component Contrastive unlearning on Graph Neighborhoods (CoGN), finds affected neighbors of the known deletion set, updating the GNN weights using a contrastive loss that pushes representations of the affected neighbors away from the deletion entities while staying close to other neighbors. The second component, AsCent DesCent de ↯ coupled (AC ↯ DC) applies the classic i.i.d. unlearning method of gradient ascent on the deletion set and gradient descent on the retain set.
Our proposed method shows promise for corrective unlearning: we not only outperform retraining from scratch, the previously assumed gold standard for this task, but also recover most of the performance of an oracle (a model trained on the complete and correct data) while discovering as few as 5% of the manipulated entities.
2Corrective Unlearning for Graph Neural Networks
We now formulate the Corrective Unlearning problem for graph-structured, non-i.i.d. data. We consider a graph 𝒢
( 𝒱 , ℰ ) , where 𝒱 and ℰ represent the constituent set of nodes and edges respectively. For each node 𝒱 𝑖 ∈ 𝒱 , there is a corresponding feature vector 𝒳 𝑖 and label 𝒴 𝑖 , with 𝒱
( 𝒳 , 𝒴 ) . Consistent with prior work in unlearning on graphs (Wu et al., 2023a; Li et al., 2024c), we focus on semi-supervised node classification using GNNs. GNNs use the message-passing mechanism, where each node aggregates features from its immediate neighbors. The effect of this aggregation process propagates through multiple successive layers, effectively expanding the receptive field of each node with network depth. This architecture inherently exploits the principle of homophily, a common property in many real-world graphs where nodes with similar features or labels are more likely to be connected than not.
While assuming homophily is extremely useful for learning representations from graph data, annotation mistakes or adversarial manipulations that create dissimilar neighborhoods or connect otherwise dissimilar nodes can easily harm the learned representations (Zügner & Günnemann, 2019). This motivates our study of post-hoc correction strategies like unlearning for GNNs. Following Goel et al. (2024), we adopt an adversarial formulation that subsumes correcting more benign mistakes.
Adversary’s Perspective. The adversary aims to reduce model accuracy on a target distribution by manipulating parts of the clean training data 𝒢 . This can be done in the following ways: (1) adding spurious edges ℰ ^ , resulting in ℰ ′
ℰ ∪ ℰ ^ ; or (2) manipulating node information, 𝒱 ′
𝑓 𝑚 ( 𝒱 ) , where 𝑓 𝑚 manipulates a subset of nodes by changing their features or labels. We define 𝑆 𝑚 as the set of manipulated entities, which can be either the manipulated subset of nodes or the added spurious edges ℰ ^ . The final manipulated graph is denoted as 𝒢 ′
( 𝒱 ′ , ℰ ′ ) .
Unlearner’s Perspective. After training, model developers may observe that desired properties like fairness and robustness are compromised in the trained model ℳ , which can be modeled as lower accuracy on some data distributions. The objective, then, is to remove the influence of the manipulated training data 𝑆 𝑚 on the affected distribution while maintaining performance on the remaining entities. By utilizing data monitoring strategies on a subset of the training data or using incorrect data detection techniques like (Northcutt et al., 2021), it may be possible to identify a part of the manipulated entities 𝑆 𝑓 ⊆ 𝑆 𝑚 . For unlearning to be feasible, 𝑆 𝑓 must be a representative subset of 𝑆 𝑚 . We only assume the type of affected entity (edges or nodes) is known to the model developer, but do not assume any knowledge about the nature of manipulation. An unlearning method 𝑈 ( ℳ , 𝑆 𝑓 , 𝒢 ′ ) is then used to mitigate the adverse effects of 𝑆 𝑚 , ideally by improving the accuracy on unseen samples from the affected distribution. An effective unlearning method should remove the impact of certain training data samples without degrading performance on the rest of the data or incurring the cost of retraining from scratch. Moreover, while Retrain was previously considered a gold standard in privacy-oriented unlearning and graph unlearning, Goel et al. (2024) showed that when the whole manipulated set is not known, retraining on the remaining data can reinforce the manipulation, implying it’s not a gold standard for corrective unlearning.
Metrics. To evaluate the performance of unlearning methods, we use the metrics proposed by Goel et al. (2024):
1. 𝐀𝐜𝐜 𝐚𝐟𝐟 : It measures the clean-label accuracy of test set samples from the affected distribution. This metric captures the method’s ability to correct the influence of the manipulated entities on unseen data through unlearning. As the affected distribution differs for each manipulation, we specify it when describing each evaluation.
2. 𝐀𝐜𝐜 𝐫𝐞𝐦 : It is defined as the accuracy of the remaining entities. This metric measures whether the unlearning maintains model performance on clean entities.
The metrics Acc aff and Acc rem were termed “Corrected Accuracy” ( Acc corr ) and “Retain Accuracy” ( Acc retain ) respectively by Goel et al. (2024). We chose alternative names to explicitly state which data distribution accuracy is measured. In Section 4, we further specify what the “affected distribution” and “remaining entities” are for the different evaluation types we study.
Goal. An ideal corrective unlearning method should have high Acc aff even when a small fraction of manipulated set ( 𝑆 𝑚 ) is identified for deletion ( 𝑆 𝑓 ) without big drops in Acc rem , all while being computationally efficient.
3Our Method: Cognac
Our proposed unlearning method, Cognac, requires access to the underlying graph 𝒢 ′ , the known set of entities to be deleted 𝑆 𝑓 , and the original model ℳ . We define 𝒱 𝑓 as the set of nodes whose influence is to be removed. For node unlearning, 𝒱 𝑓
𝑆 𝑓 ; for edge unlearning, 𝒱 𝑓 is the set of vertices connected to the edge set to be deleted. Manipulated data has two main adverse effects on the trained GNN: 1) Message passing can propagate the influence of the manipulated entities 𝑆 𝑚 on their neighborhood, and 2) The layers learn transformations to fit potentially wrong labels in 𝑆 𝑚 . Mitigating the effects of attacks first requires analysis of the impact of such attacks. The most general form of attack is Interclass Confusion (Goel et al., 2022), which we show entangles the representations of two classes, below.
Theorem 3.1.
Let 𝐺
( 𝑉 , 𝐸 , 𝑋 ) be a graph with node set 𝑉 , edge set 𝐸 , and features 𝑋 . Let 𝐶 1 , 𝐶 2 ⊂ 𝑉 be two distinct classes with ground-truth labels 𝑦 𝑖 ∈ { 𝐶 1 , 𝐶 2 } . Suppose an Interclass Confusion (IC) attack is applied.
Let 𝜙 𝑀 ( 𝐶 1 ) , 𝜙 𝑀 ( 𝐶 2 ) ∈ ℝ 𝑑 denote the mean embeddings of 𝐶 1 and 𝐶 2 , and 𝒟 ( 𝜙 𝑀 ( 𝐶 1 ) , 𝜙 𝑀 ( 𝐶 2 ) ) be the Wasserstein-2 distance between their embedding distributions.
Then, there exists a degradation term Δ
0 such that:
𝔼 [ 𝒟 ( 𝜙 𝑀 ( 𝐶 1 ) , 𝜙 𝑀 ( 𝐶 2 ) ) ]
≤ 𝔼 [ 𝒟 ( 𝜙 𝑀 clean ( 𝐶 1 ) , 𝜙 𝑀 clean ( 𝐶 2 ) ) ] − Δ
Complete list of assumptions and attack details in Appendix A.3. We tackle these two problems using separate components - CoGN and AC ↯ DC.
3.1Removing Effects on Neighbors with CoGN
The first question we address is: How can we remove the influence of manipulated entities on their neighboring nodes? First, this requires us to identify the nodes affected by the manipulations ( 𝒱 aff ) and then mitigate the influence on their representations. Identifying affected nodes is challenging, as the impact of message passing from manipulated entities 𝑆 𝑚 depends on the interference from messages of other neighboring nodes. Therefore, we use an empirical estimation to identify the affected nodes from each entity in the deletion set. On these nodes, we then perform contrastive unlearning, simultaneously pushing the representations of the affected nodes away from nodes in 𝒱 𝑓 while keeping them close to other nodes in their neighborhood. We call this component Contrastive unlearning on Graph Neighborhoods (CoGN), formalized below.
3.1.1Affected Node Identification
To first identify the affected samples,
𝒱
aff
, we first observe that any manipulated node
𝑠
∈
𝑉
𝑓
can only affect the representation of any other node
𝑣
∈
𝒱
′
𝑉
𝑓
in
𝒢
′
only if
𝑠
is part of the receptive field of
𝑣
, a widely known result for GNNS. Formally,
Lemma 3.2.
Let 𝐺
( 𝑉 , 𝐸 ) be an undirected graph, and let 𝒩 ( 𝑣 ) denote the 1-hop neighborhood of node 𝑣 . For a node 𝑠 ∈ 𝑉 , let 𝒩 𝑛 ( 𝑠 ) denote the 𝑛 -hop neighborhood of 𝑠 , defined recursively as:
𝒩 𝑛 ( 𝑠 )
{
{
𝑠
}
if
𝑛
0 ,
⋃ 𝑣 ∈ 𝒩 𝑛 − 1 ( 𝑠 ) 𝒩 ( 𝑣 )
if 𝑛
0 .
In an 𝑛 -layer GNN, the representation 𝑧 𝑠 of node 𝑠 can affect the representations 𝑧 𝑣 of nodes 𝑣 only if 𝑣 ∈ 𝒩 𝑛 ( 𝑠 ) . For any 𝑣 ∉ 𝒩 𝑛 ( 𝑠 ) , 𝑧 𝑣 is independent of message passing effects of 𝑧 𝑠 .
The proof is provided in Appendix A.1. The main takeaway is that manipulations only propagate within an 𝑛 -hop neighborhood of poisoned nodes. This locality drastically reduces the search space for identifying affected nodes.
The second observation is that not all nodes in the 𝑛 -hop neighborhood of the manipulated nodes may be affected enough by the attack, as some nodes are more robust to the perturbations than others (Gosch et al., 2023; Arun et al., 2023). To find the most affected nodes, we employ a cheap and simple heuristic. We invert the features of 𝑣 ∈ 𝑉 𝑓 and select neighboring nodes where final output logits are changed the most. Formally, the inversion is performed by the transformation 1 → − 𝒳 𝑣 , ∀ 𝑣 ∈ 𝒱 𝑓 , leading to a new feature matrix 𝜒 ′ , where 𝒳 𝑣 represents a one-hot-encoding vector. We then compute the difference in the original output logits ℳ ( 𝜒 ) , and those obtained by on the new feature matrix, ℳ ( 𝜒 ′ ) given by: Δ 𝜒
| ℳ ( 𝜒 ′ ) − ℳ ( 𝜒 ) | . The top 𝑘 % nodes with the most change, Δ 𝜒 , are selected as the affected set of entities 𝒱 aff . A conceptual example illustrating the workings of Affected Node Identification is presented in Figure 2. Appendix E.2 varies our design choices, confirming that we retain the same performance as using the entire 𝑛 -hop neighborhood while being more efficient (Figure 9). Our method works robustly even if the original GNN was under-trained (Figure 8). Further, in Table 7 we also ablate the heuristic for identifying affected nodes against Cognac using MEGU’s sampling technique. We observe that our heuristic delivers over 25 % higher and is 8 x faster.
Figure 2:A toy example detailing Affected Node Identification. (1) We first invert the features of the known manipulated node. (2) We perform two forward passes through the GNN, one with the original feature vector and one with the inverted feature vector. (3) We compute the difference in output logits from both cases and take the top 2 nodes with the largest logit change, which in this case are nodes E and B. 3.1.2Contrastive Unlearning
To remove the influence of the deletion set
𝒱
𝑓
on the affected nodes
𝒱
aff
identified in the previous step, we must ensure that our model maps the hidden representations of nodes in
𝒱
aff
far away from that of nodes in
𝒱
𝑓
(Goel et al., 2024). However, satisfying this property alone will lead to unrestricted separation and damage the quality of learned representations. To achieve this balance and restore homophily (Ma et al., 2022), we constrain
𝒱
aff
to stay close to its unaffected neighbors
𝒩
𝒱
aff
𝒱
𝑓
𝒱 𝑝 𝑜 𝑠 , ensuring alignment with the local structure while distancing from 𝒱 𝑓 . Formally, we can state this as the following optimization problem over the parameters of a GNN, which is necessary for corrective unlearning,
max 𝜃 ( 𝔼 𝑣 ∈ 𝒱 aff , 𝑝 ∈ 𝒱 𝑝 𝑜 𝑠 [ 𝐳 𝑣 ⊤ 𝐳 𝑝 ] − 𝔼 𝑣 ∈ 𝒱 aff , 𝑛 ∈ 𝒱 𝑓 [ 𝐳 𝑣 ⊤ 𝐳 𝑛 ] )
max 𝜃 ( 𝔼 𝑣 ∈ 𝒱 aff , 𝑝 ∈ 𝒱 𝑝 𝑜 𝑠 , 𝑛 ∈ 𝒱 𝑓 [ 𝐳 𝑣 ⊤ 𝐳 𝑝 − 𝐳 𝑣 ⊤ 𝐳 𝑛 ] )
(1)
Equation 1 offers a direct way to enforce separation between positive and negative pairs, but it has notable shortcomings. In particular, it does not sufficiently penalize small margins, as it only considers the raw similarity difference without emphasizing cases where 𝑧 𝑣 𝑇 𝑧 𝑝 and 𝑧 𝑣 𝑇 𝑧 𝑛 are nearly equal. This can lead to weak separation and reduced robustness, especially when positive and negative embeddings are closely aligned. Additionally, the lack of non-linear scaling creates an uneven optimization landscape, increasing the risk of convergence issues. To overcome these limitations, we use a log-based loss function that applies non-linear sigmoid terms, providing stronger penalization for small margins and a probabilistic similarity interpretation (Hamilton et al., 2017). This enhances the separation between positive and negative pairs while ensuring smoother gradients for more stable optimization (Lee et al., 2024). Formally,
ℒ 𝑣 , 𝑝 , 𝑛
− log ( 𝜎 ( 𝑧 𝑣 𝑇 𝑧 𝑝 ) ) − log ( 𝜎 ( − 𝑧 𝑣 𝑇 𝑧 𝑛 ) )
(2)
Under assumptions of differentiability, convexity, and bounded gradients, we can guarantee that optimizing this function achieves a better separation between the positive and negative dot products equivalent to Equation 1.
Theorem 3.3.
Let 𝜃 ∈ ℝ 𝑑 parameterize a GNN generating embeddings 𝑧 𝑣 , 𝑧 𝑝 , 𝑧 𝑛 ∈ ℝ 𝑘 for triplets ( 𝑣 , 𝑝 , 𝑛 ) . Let ℒ ( 𝜃 )
− 𝔼 [ log 𝜎 ( 𝑧 𝑣 ⊤ 𝑧 𝑝 ) + log 𝜎 ( − 𝑧 𝑣 ⊤ 𝑧 𝑛 ) ] and 𝑆 ( 𝜃 )
𝔼 [ 𝑧 𝑣 ⊤ 𝑧 𝑝 − 𝑧 𝑣 ⊤ 𝑧 𝑛 ] .
Then under the assumptions stated above, gradient descent on ℒ ( 𝜃 ) with step size 𝜂 ≤ 1 𝐿 (where 𝐿 is the Lipschitz constant of ∇ 𝜃 ℒ ) guarantees:
𝑆 ( 𝜃 𝑡 + 1 ) ≥ 𝑆 ( 𝜃 𝑡 ) ∀ 𝑡 ≥ 0 ,
and at convergence, 𝑆 ( 𝜃 ∗ )
𝑆 ( 𝜃 0 ) .
The proof is presented in Appendix A.2. We also empirically validate this theorem for non-linear GNNs, with Figure 3 showing on the Cora dataset that the separation between the positive dot product and negative dot product monotonically increases as the loss converges.
3.2Unlearning Old Labels with AC ↯ DC
Next, we ask: Can we undo the effect of the task loss ℒ task explicitly learning to fit the node representations of 𝑆 𝑚 to potentially wrong labels? We do this by performing gradient ascent on 𝑆 𝑓 , which non-directionally maximizes the training loss with respect to the old labels. Ascent alone aggressively leads to arbitrary forgetting of useful information, so we counterbalance it by alternating with steps that minimize the task loss on the remaining data. More precisely, we perform gradient ascent on 𝒱 𝑓 and gradient descent on 𝒱 𝑟 , iteratively on the original GNN ℳ .
ℒ 𝑎
− ℒ task ( 𝒱 𝑓 ) , ℒ 𝑑
ℒ task ( 𝒱 𝑟 )
(3)
While variants of ascent on 𝑆 𝑓 and descent on remaining data have been studied for image classification (Kurmanji et al., 2023) and language models (Yao et al., 2024), we find a specific optimization strategy useful to achieve corrective unlearning on graphs. The challenge arises when 𝑆 𝑓 ⊂ 𝑆 𝑚 , as the remaining data could still contain manipulated entities, which we aim to avoid reinforcing. However, in realistic scenarios, the manipulated entities 𝑆 𝑚 typically are a small fraction of the training data, allowing us to mitigate their impact through ascent on the representative subset 𝑆 𝑓 .
Figure 3:Empirical convergence of CoGN. The average positive and negative dot products across samples increase and decrease, respectively, over epochs, resulting in overall convergence.
This requires a careful balance between ascent and descent, which we can achieve by using two different optimizers and starting learning rates for these steps. This insight is similar to prior work in Generative Adversarial Networks (GANs) (Heusel et al., 2017). The starting learning rates for both ascent and descent are hyperparameters to be tuned, and usually, we find that a lower learning rate for ascent leads to better results. Thus, we call this component Ascent Descent de ↯ coupled. Convergence of this formulation has been shown in previous works (Kurmanji et al., 2023). We show the empirical convergence of AC ↯ DC in Figure 7 present in Appendix E.1.
For our final method Cognac, we alternate steps of CoGN, which fixes representations of affected neighborhood nodes, and AC ↯ DC, which unlearns potentially wrong labels introduced by 𝑆 𝑚 . We also perform ablations in Section 5 (Table 3), showing that the individual components CoGN and AC ↯ DC alone perform notably worse than our final method, suggesting the necessity of both components.
4Experimental Setup 4.1Benchmarking Details
We now describe design choices made for benchmarking, first specifying the datasets and architectures, and then outlining how to ensure a fair comparison between methods.
Models and Datasets. We report results using the Graph Convolutional Network (GCN) (Kipf & Welling, 2017) architecture and evaluate the methods on ten benchmark datasets used in prior literature: Cora, PubMed, DBLP, Coauthor CS, Coauthor Physics, Amazon Photos (Photos), Amazon Computers (Computers), CiteSeer (Cheng et al., 2023; Li et al., 2024c), and additionally CoraFull (Bojchevski & Günnemann, 2018) and OGB-arXiv (Wang et al., 2020). In Appendix D.1, we provide additional results on Graph Attention Network (GAT) (Veličković et al., 2018). For each dataset, we extract the largest connected subgraph for our experiments. Dataset details, including the number of classes, nodes, edges, and entities manipulated, are provided in Table 1 and Appendix B.
Hyperparameter Tuning. Ensuring a fair comparison of unlearning methods can be tricky, as there are multiple desiderata: unlearning, maintaining utility, and computational efficiency, and hyperparameter tuning of the methods can particularly affect results on GNNs. We describe our efforts towards this in Appendix sections F.1 and F.2.
Table 1:Dataset statistics. In our evaluations, we include citation networks, co-author networks, and co-purchase networks. The number of nodes and edges reported here refers to the entire dataset. From this, we use a 60/20/20 split for train/validation/test. The manipulation statistics are presented in Table 4.
Dataset Classes Nodes Edges CoraFull 70
18 , 800
125 , 370
Cora 7
2 , 485
10 , 138
CiteSeer 6
2 , 120
7 , 358
DBLP 4
16 , 191
103 , 826
PubMed 3
19 , 717
88 , 648
OGB-arXiv 40
169 , 343
1 , 166 , 243
CS 15
18 , 333
163 , 788
Physics 5
34 , 493
495 , 924
Photos 8
7 , 487
238 , 086
Computers 10
13 , 381
491 , 556 4.2Evaluations
Given a fixed budget of samples to manipulate, ideal corrective unlearning evaluations should maximize the deterioration of model performance on the affected distribution, creating a wide gap between clean and poisoned model performance to measure the progress of the unlearning method. We thus evaluate unlearning in the context of attacks that are not constrained by stealthiness. Lingam et al. (2024) show that binary label flip manipulation attacks, where a fraction of labels are swapped between two chosen classes, are stronger than multi-class manipulations, theoretically and empirically, on GNNs. Building on this, we use two targeted attacks to evaluate corrective unlearning on graph data. Additionally, for completeness we also present results on a Feature Poisoning Attack, analogous to patch-based backdoor attacks in image poisoning (Gu et al., 2017), in Appendix D.3.
Spurious Edge Addition. Prior GNN unlearning works (Wu et al., 2023a; Li et al., 2024c) have evaluated adversarial edge attacks, but in an untargeted setting, making their evaluations weak. We, instead, simulate targeted adversarial edge insertions between nodes of two classes, violating homophily assumptions and entangling their representations. This models attacks like fake social connections (Bojchevski & Günnemann, 2019a) or knowledge graph manipulations (Xi et al., 2023; Zhang et al., 2019; Zhao et al., 2024). Unlearning aims to recover accuracy on the targeted classes Acc aff while preserving performance on others Acc rem .
Label Manipulation. We implement the Interclass Confusion (IC) Test (Goel et al., 2022), by systematically swapping labels between two targeted classes to entangle their representations. Once again, the unlearning goal is to improve Acc aff on the two targeted classes, while preserving performance, Acc rem , on the remaining classes.
4.3Baselines
We evaluate four popular graph unlearning methods and adapt one popular i.i.d. unlearning method for graphs. For reference, we also report results for the Original model, Retrain, which trains a new model without 𝑆 𝑓 , and Finetune, which continues training the poisoned model on data without 𝑆 𝑓 for additional epochs. Following Goel et al. (2024), we find that retraining from scratch is not the gold standard in corrective unlearning. Thus, we introduce Oracle, trained on the whole training set without manipulations, indicating an upper bound on what can be achieved. The Oracle has correct labels for the unlearning entities, information that the unlearning methods cannot access.
Existing Unlearning Methods. We choose five methods as baselines where unlearning incorrect data explicitly motivates the technique. (1) GNNDelete (Cheng et al., 2023) adds a deletion operator after each GNN layer and trains them using a loss function to randomize the prediction probabilities of deleted edges while preserving their local neighborhood representation, keeping the original GNN weights unchanged. (2) GIF (Wu et al., 2023a) draws from a closed-form solution for linear GNNs to measure the structural influence of deleted entities on their neighbors. Then, they provide estimated GNN parameter changes for unlearning using the inverse Hessian of the loss function. (3) MEGU (Li et al., 2024c) finds the highly influenced neighborhood (HIN) of the unlearning entities and removes their influence over the HIN while maintaining predictive performance and forgetting the deletion set using a combination of losses. (4) UtU (Tan et al., 2024) proposes zero-cost edge-unlearning by removing the edges to be deleted during inference for blocking message propagation from nodes linked to these edges. Finally, we include a popular unlearning method studied in i.i.d. classification settings. (5) SCRUB (Kurmanji et al., 2023) employs a teacher-student framework with alternate steps of distillation away from the forget set and towards the retain set. For edge unlearning, we use SCRUB by taking the nodes connected to spuriously added edges as the forget set and the rest as the retain set.
Figure 4:Corrective Unlearning Results. We report the accuracy on the affected classes Acc aff across different fractions of the manipulation set known for deletion ( 𝑆 𝑓 / 𝑆 𝑚 ) . Baseline methods perform poorly, except for GNNDelete, which achieves reasonable unlearning performance in some settings. AC ↯ DC and Finetune, despite not being graph-specific, perform much better. Cognac, which adds graph awareness to AC ↯ DC, archives SOTA across datasets and corrective fractions, unlearning the effect of the manipulation with just 5% of the manipulation set known. 5Results & Discussion
We now report our main results comparing our method to existing methods across the manipulation types and datasets. Detailed method ablations and analyses of what can be achieved in this setting are reported in Appendix E.4. We also present consistent results on the GAT architecture in Appendix D.1. We show the robustness of Cognac to large 𝑆 𝑓 sizes, showing strong performance even with 38.96 % of total training nodes in 𝑆 𝑓 for the PubMed dataset.
Table 2:Accuracy averaged across 𝑆 𝑓 / 𝑆 𝑚 on remaining distribution relative to the Original model. We find prior methods, especially GNNDelete, lead to large drops in Acc rem , while our methods Cognac and AC ↯ DC minimize the loss in Acc rem . Method CS Cora Edge Label Edge Label Original \cellcolor[HTML]FFFFFF 90.6 \cellcolor[HTML]FFFFFF 89.6 \cellcolor[HTML]FFFFFF 61.2 \cellcolor[HTML]FFFFFF 61.4
Oracle \cellcolor[HTML]FFFFFF − 0.1 \cellcolor[HTML]FFFFFF + 0.5 \cellcolor[HTML]FFFEFE − 0.7 \cellcolor[HTML]FFF8F8 − 3.0
Retrain \cellcolor[HTML]FFFCFC − 1.4 \cellcolor[HTML]FFFFFF − 0.1 \cellcolor[HTML]FFF9F9 − 2.4 \cellcolor[HTML]FFF1F1 − 5.7
Cognac \cellcolor[HTML]FFFCFC − 1.4 \cellcolor[HTML]FFFEFE − 0.4 \cellcolor[HTML]FFF4F4 − 4.5 \cellcolor[HTML]FFF8F8 − 2.9
AC ↯ DC \cellcolor[HTML]FFFEFE − 0.7 \cellcolor[HTML]FFFDFD − 0.8 \cellcolor[HTML]FFFAFA − 2.2 \cellcolor[HTML]FFFEFE − 0.8
GNNDelete \cellcolor[HTML]FFF0F0 − 6.0 \cellcolor[HTML]FFFBFB − 1.9 \cellcolor[HTML]FFE4E4 − 10.9 \cellcolor[HTML]FFEAEA − 8.4
GIF \cellcolor[HTML]FFFBFB − 1.6 \cellcolor[HTML]FFFDFD − 0.9 \cellcolor[HTML]FFF5F5 − 4.2 \cellcolor[HTML]FFFEFE − 0.6
MEGU \cellcolor[HTML]FFFEFE − 0.5 \cellcolor[HTML]FFF8F8 − 2.8 \cellcolor[HTML]FFFFFF + 0.0 \cellcolor[HTML]FFEFEF − 6.5
UtU \cellcolor[HTML]FFFFFF + 0.0 \cellcolor[HTML]FFFFFF + 0.0 \cellcolor[HTML]FFFFFF + 0.0 \cellcolor[HTML]FFFFFF + 0.0
SCRUB \cellcolor[HTML]FFFDFD − 0.9 \cellcolor[HTML]FFFEFE − 0.7 \cellcolor[HTML]FFFFFF + 0.0 \cellcolor[HTML]FFF3F3 − 4.8
Figure 4 shows unlearning performance on the test set for manipulated classes ( Acc aff ) upon varying the fraction of the manipulation set known for unlearning ( 𝑆 𝑓 / 𝑆 𝑚 ) . Table 2 accompanies this, showing side-effects on utility.
Existing unlearning methods perform poorly even when | 𝑆 𝑓 |
| 𝑆 𝑚 | . Observing the rightmost points in Figure 4, we can see across manipulation types and datasets that existing methods fail to improve Acc aff even when the whole manipulation set is known. UtU fails to unlearn the effects of either of the attacks, as simply unlinking on the forward pass does not sufficiently counteract the influence on neighbors and weights. Both SCRUB and MEGU use a KL Divergence Loss term to keep predictions on the remaining data close to the original model, which could be detrimental when done on unidentified manipulation set entities and other affected neighbors. Even though MEGU and GIF were evaluated on removing adversarial edges and GNNDelete also mentioned incorrect data as one of its key applications, they failed to recover performance when presented with targeted data manipulation. Interestingly, despite extensive hyperparameter searches, they are beaten by methods with no special graph components: the best baselines are AC ↯ DC and naive finetuning (Finetune) on the retain set: both achieve a performance similar to retraining in some of the evaluations.
Corrective unlearning methods must perform better than Retrain. While both AC ↯ DC and Finetune match Retrain, all of them still fall far behind the Oracle model’s performance in most evaluations, and are not consistent enough. This scope for improvement motivates the design of our method which performs well across attacks, datasets, and fractions identified for deletion.
Cognac beats Retrain, and can sometimes match the Oracle’s performance. We observe that Cognac consistently achieves state-of-the-art performance across all datasets and manipulations, convincingly and consistently beating existing graph methods, and often exceeds Retrain. Notably, Cognac occasionally even surpasses Oracle, our introduced gold standard, on multiple datasets and identified fractions - despite having access to less data (and unknown manipulated samples) than Oracle. In Table 3, we show how both components of Cognac complement each other. CoGN alone does not improve Acc aff over the original model, showing that while it effectively moves affected nodes, it lacks signals from labels. Conversely, AC ↯ DC alone achieves better performance than CoGN, but is still far from matching the Oracle. AC ↯ DC weakens incorrect learning signals and preserves task-relevant representations, while CoGN steers affected nodes away from manipulated ones. This highlights the effectiveness of our contrastive approach in achieving robust unlearning.
Cognac performs strongly even with only 𝟓 % of 𝐒 𝐦 known. Cognac effectively recovers most of the accuracy on the affected distribution even when only 5 % of the manipulated set is known. Notably, it outperforms Retrain in realistic scenarios where 𝑆 𝑚 is only partially known. We attribute this to Affected Neighborhood Identification, which leverages graph structure to infer manipulated nodes and edges beyond those explicitly identified. Additionally, CoGN plays a crucial role by pushing influenced neighbors away from the deletion set and aligning them with their unaffected neighbors, thereby correcting representations even for unknown manipulated samples.
Table 3:Ablating both components of Cognac across datasets. CoGN alone does not improve Acc aff over the original model. Conversely, AC ↯ DC alone achieves better performance than CoGN, but is still far from matching the Oracle. These results show how both components are integral to Cognac’s success. Method CS Cora
Acc aff
Acc rem
Acc aff
Acc rem
Original 44.3
89.8
51.4
60.6
Oracle 90.1
90.0
71.1
60.7
CoGN 47.2
89.8
42.1
60.5
AC ↯ DC 67.2
86.6
59.9
61.3
Cognac 79.3
82.3
75.5
56.6
- Cognac scales well to significantly larger datasets. We evaluate Cognac on OGB-Arxiv, a significantly larger dataset than our other benchmarks with 170 , 000 nodes and over 1 M edges, and observe that it continues to achieve strong performance. Despite the increased scale, Cognac maintains a substantial lead over Retrain across different fractions of 𝑆 𝑚 (Figure 5 (a)), while the baseline methods fail to improve Acc aff over the original model. Even as dataset size grows, Cognac effectively retains its advantage, demonstrating its scalability and robustness.
Figure 5:(a) Results on a larger dataset, OGB-Arxiv, and (b) Visualization of hidden layer embeddings after unlearning on CS dataset for node unlearning. (a) On the OGB-Arxiv dataset, Cognac outperforms retraining from scratch by more than 10%, while most baselines fail to achieve any performance gains beyond the Original model. (b) The affected distribution embeddings (highlighted by red and blue) are fully entangled in the original trained model, while after unlearning with Cognac the embeddings are well separated and clustered, matching Oracle.
Overall, our work makes progress on the problem of corrective unlearning in graph neural networks with remarkably minimal training signal: we achieve strong unlearning with the knowledge of as little as ( 5 % ) of the manipulation set 𝑆 𝑚 . The visualization of the hidden GNN layer embeddings after unlearning (Figure 5 (b)) shows that Cognac, like the Oracle model, achieves clear clustering of the manipulated class data points, validating our theoretical results shown in Section 3.1.
6Related Work
Graph-based attacks, such as Sybil (Douceur, 2002) and link spam farms (Wu & Davison, 2005), have long affected the integrity of social networks and search engines. Recent works reveal that even state-of-the-art GNN architectures are vulnerable to simple attacks on the trained models, which either manipulate existing edges and nodes or inject new adversarial nodes (Sun et al., 2019; Dai et al., 2018; Zügner & Günnemann, 2019; Geisler et al., 2024). Parallelly, works have characterized the influence of specific nodes and edges that can guide such attacks (Chen et al., 2023). One strategy to mitigate the influence of such attacks is robust pretraining, such as using adversarial training (Yuan et al., 2024; Zhang et al., 2023). Post-hoc interventions like unlearning act as a complementary layer of defense, helping model developers when attacks slip through and affect a trained model.
Removing the impact of manipulated entities begins with their identification (Brodley & Friedl, 1999), for which multiple strategies exist like data attribution (Ilyas et al., 2022), adversarial detection, and automated or human-in-the-loop anomaly detection (Northcutt et al., 2021). While approaches like model debiasing and concept erasure (Fan et al., 2024; Belrose et al., 2023) can remove effects of identified manipulations post-training, they require knowledge of the manipulation. In contrast, unlearning methods are particularly valuable in adversarial settings where corruption effects may be deliberately obfuscated and impact multiple model behaviors simultaneously (Paleka & Sanyal, 2023).
Recently, machine unlearning has received newfound attention beyond privacy applications (Pawelczyk et al., 2024; Schoepf et al., 2024; Li et al., 2024a, b). Goel et al. (2024) demonstrated the distinction between the Corrective and Privacy-oriented unlearning settings for i.i.d. classification tasks, emphasizing challenges when not all manipulated data is identified for unlearning.
Exact Unlearning arose in privacy applications, offering guaranteed removal of data influence through selective retraining (Chen et al., 2022b, a; Bourtoule et al., 2021). While perfect guarantees are valuable for privacy, the exponential cost of sequential deletions (Warnecke et al., 2023) becomes impractical as unlearning expands to broader challenges like model correction and debiasing (Pawelczyk et al., 2024; Schoepf et al., 2024; Li et al., 2024a). This has driven the development of Inexact Unlearning methods that balance effectiveness with scalability, using either theoretical bounds for simple models (Chien et al., 2022; Wu et al., 2023b) or empirical validation for deep networks (Wu et al., 2023a; Cheng et al., 2023; Li et al., 2024c). The recent formalization of Corrective Unlearning (Goel et al., 2024) opened a crucial new direction: removing corruption effects with only partial identification of manipulated data. We tackle this challenge in graphs, where the non-i.i.d. nature of data (Said et al., 2023) results in the graph elements exerting a strong influence on other elements in their neighborhood.
7Limitations and Conclusion
Our work addresses corrective unlearning for GNNs, focusing on removing the effects of manipulated training data when only a fraction is identified. As our method relies on the homophily assumption and is evaluated primarily on homophilic datasets like previous works (Li et al., 2024c), it is left for future works to expand this to heterophilic datasets. While our method does not guarantee successful unlearning against arbitrary real-world attacks, our experiments suffice to show that existing unlearning methods struggle even with complete knowledge of manipulations, which is an unrealistic scenario.
Cognac pushes the frontiers of unlearning by consistently beating retraining-from-scratch, and nearly matching the performance of a strong oracle on unlearning class confusion manipulations with access to as little as 5 % of the manipulated data. We hope this sparks interesting future work on developing stronger evaluations and theoretical understanding for graph corrective unlearning in the GNN Robustness and Machine Unlearning community.
Acknowledgements
Varshita Kolipaka is grateful to be funded by IHUB-Data. Arvindh Arun was funded by the CHIPS Joint Undertaking (JU) under grant agreement No. 101140087 (SMARTY), and by the German Federal Ministry of Education and Research (BMBF) under the sub-project with the funding number 16MEE0444, and acknowledges support from the International Max Planck Research School for Intelligent Systems (IMPRS-IS) and the European Laboratory for Learning and Intelligent Systems (ELLIS) PhD programs.
The authors extend their appreciation to the members of the Precog research group, Shashwat Singh, Karuna Chandra, Pratyaksh Gautam, Prashant Kodali, and Makarand Tapaswi for helpful discussions and feedback.
Impact Statement
The increasing adoption of Graph Neural Networks (GNNs) in real-world applications raises concerns about fairness and safety, particularly in settings where biased data propagates through message passing or incorrect information influences high-stakes decisions. From a fairness perspective, Cognac can be potentially used to mitigate the impact of biased social connections in recommender systems and hiring networks, where unfair edges or annotations may reinforce discriminatory patterns. From a safety standpoint, GNNs are increasingly used in drug discovery and biomedical applications, where incorrect relationships between molecular compounds or erroneous biological interactions could lead to misleading predictions. Corrective unlearning can help remove the influence of faulty training data or adversarially inserted edges, improving the reliability of GNN-based scientific models without compromising efficiency.
References Akiba et al. (2019) ↑ Akiba, T., Sano, S., Yanase, T., Ohta, T., and Koyama, M.Optuna: A next-generation hyperparameter optimization framework.In Proceedings of the 25th ACM SIGKDD international conference on knowledge discovery & data mining, pp. 2623–2631, 2019. Arun et al. (2023) ↑ Arun, A., Aanegola, A., Agrawal, A., Narayanam, R., and Kumaraguru, P.CAFIN: Centrality Aware Fairness inducing IN-processing for unsupervised representation learning on graphs.In Proceedings of the 26th European Conference on Artificial Intelligence, 2023.doi: 10.3233/faia230259.URL http://dx.doi.org/10.3233/FAIA230259. Arun et al. (2025) ↑ Arun, A., Kumar, S., Nayyeri, M., Xiong, B., Kumaraguru, P., Vergari, A., and Staab, S.Semma: A semantic aware knowledge graph foundation model, 2025.URL https://arxiv.org/abs/2505.20422. Belrose et al. (2023) ↑ Belrose, N., Schneider-Joseph, D., Ravfogel, S., Cotterell, R., Raff, E., and Biderman, S.LEACE: Perfect linear concept erasure in closed form.In Thirty-seventh Conference on Neural Information Processing Systems, 2023.URL https://openreview.net/forum?id=awIpKpwTwF. Bojchevski & Günnemann (2018) ↑ Bojchevski, A. and Günnemann, S.Deep gaussian embedding of graphs: Unsupervised inductive learning via ranking.In International Conference on Learning Representations, 2018. Bojchevski & Günnemann (2019a) ↑ Bojchevski, A. and Günnemann, S.Adversarial attacks on node embeddings via graph poisoning.In International Conference on Machine Learning, pp. 695–704. PMLR, 2019a. Bojchevski & Günnemann (2019b) ↑ Bojchevski, A. and Günnemann, S.Certifiable robustness to graph perturbations.Advances in Neural Information Processing Systems, 32, 2019b. Bourtoule et al. (2021) ↑ Bourtoule, L., Chandrasekaran, V., Choquette-Choo, C. A., Jia, H., Travers, A., Zhang, B., Lie, D., and Papernot, N.Machine unlearning.In IEEE S&P, 2021. Brodley & Friedl (1999) ↑ Brodley, C. E. and Friedl, M. A.Identifying mislabeled training data.Journal of artificial intelligence research, 11:131–167, 1999. Chen et al. (2022a) ↑ Chen, C., Sun, F., Zhang, M., and Ding, B.Recommendation unlearning.In Proceedings of the ACM Web Conference 2022, pp. 2768–2777, 2022a. Chen et al. (2022b) ↑ Chen, M., Zhang, Z., Wang, T., Backes, M., Humbert, M., and Zhang, Y.Graph unlearning.In Proceedings of the 2022 ACM SIGSAC conference on computer and communications security, pp. 499–513, 2022b. Chen et al. (2023) ↑ Chen, Z., Li, P., Liu, H., and Hong, P.Characterizing the influence of graph elements.In The Eleventh International Conference on Learning Representations, 2023.URL https://openreview.net/forum?id=51GXyzOKOp. Cheng et al. (2023) ↑ Cheng, J., Dasoulas, G., He, H., Agarwal, C., and Zitnik, M.GNNDelete: A general unlearning strategy for graph neural networks.In International Conference on Learning Representations, 2023.URL https://openreview.net/forum?id=X9yCkmT5Qrl. Chien et al. (2022) ↑ Chien, E., Pan, C., and Milenkovic, O.Efficient model updates for approximate unlearning of graph-structured data.In The Eleventh International Conference on Learning Representations, 2022. Dai et al. (2018) ↑ Dai, H., Li, H., Tian, T., Huang, X., Wang, L., Zhu, J., and Song, L.Adversarial attack on graph structured data.In Dy, J. and Krause, A. (eds.), Proceedings of the 35th International Conference on Machine Learning, volume 80 of Proceedings of Machine Learning Research, pp. 1115–1124. PMLR, 10–15 Jul 2018. Douceur (2002) ↑ Douceur, J. R.The sybil attack.In International workshop on peer-to-peer systems, pp. 251–260. Springer, 2002. Fan et al. (2024) ↑ Fan, S., Wang, X., Mo, Y., Shi, C., and Tang, J.Debiasing graph neural networks via learning disentangled causal substructure.In Proceedings of the 36th International Conference on Neural Information Processing Systems, NIPS ’22, Red Hook, NY, USA, 2024. Curran Associates Inc.ISBN 9781713871088. Geisler et al. (2024) ↑ Geisler, S., Schmidt, T., Şirin, H., Zügner, D., Bojchevski, A., and Günnemann, S.Robustness of graph neural networks at scale.In Proceedings of the 35th International Conference on Neural Information Processing Systems, NIPS ’21, Red Hook, NY, USA, 2024. Curran Associates Inc.ISBN 9781713845393. Goel et al. (2022) ↑ Goel, S., Prabhu, A., Sanyal, A., Lim, S.-N., Torr, P., and Kumaraguru, P.Towards adversarial evaluations for inexact machine unlearning.arXiv preprint arXiv:2201.06640, 2022. Goel et al. (2024) ↑ Goel, S., Prabhu, A., Torr, P., Kumaraguru, P., and Sanyal, A.Corrective machine unlearning.Transactions on Machine Learning Research, 2024.ISSN 2835-8856.URL https://openreview.net/forum?id=v8enu4jP9B. Gosch et al. (2023) ↑ Gosch, L., Sturm, D., Geisler, S., and Günnemann, S.Revisiting robustness in graph machine learning.In The Eleventh International Conference on Learning Representations, 2023.URL https://openreview.net/forum?id=h1o7Ry9Zctm. Gu et al. (2017) ↑ Gu, T., Dolan-Gavitt, B., and Garg, S.Badnets: Identifying vulnerabilities in the machine learning model supply chain.arXiv preprint arXiv:1708.06733, 2017. Günnemann (2022) ↑ Günnemann, S.Graph neural networks: Adversarial robustness.Graph neural networks: foundations, frontiers, and applications, pp. 149–176, 2022. Hamilton et al. (2017) ↑ Hamilton, W., Ying, Z., and Leskovec, J.Inductive representation learning on large graphs.Advances in neural information processing systems, 30, 2017. Heusel et al. (2017) ↑ Heusel, M., Ramsauer, H., Unterthiner, T., Nessler, B., and Hochreiter, S.Gans trained by a two time-scale update rule converge to a local nash equilibrium.Advances in neural information processing systems, 30, 2017. Ilyas et al. (2022) ↑ Ilyas, A., Park, S. M., Engstrom, L., Leclerc, G., and Madry, A.Datamodels: Understanding predictions with data and data with predictions.In Chaudhuri, K., Jegelka, S., Song, L., Szepesvari, C., Niu, G., and Sabato, S. (eds.), Proceedings of the 39th International Conference on Machine Learning, volume 162 of Proceedings of Machine Learning Research, pp. 9525–9587. PMLR, 17–23 Jul 2022.URL https://proceedings.mlr.press/v162/ilyas22a.html. Jin et al. (2020) ↑ Jin, W., Ma, Y., Liu, X., Tang, X., Wang, S., and Tang, J.Graph structure learning for robust graph neural networks.In Proceedings of the 26th ACM SIGKDD international conference on knowledge discovery & data mining, pp. 66–74, 2020. Kipf & Welling (2017) ↑ Kipf, T. N. and Welling, M.Semi-supervised classification with graph convolutional networks, 2017. Konstantinov & Lampert (2022) ↑ Konstantinov, N. H. and Lampert, C.Fairness-aware pac learning from corrupted data.JMLR, 2022. Kurmanji et al. (2023) ↑ Kurmanji, M., Triantafillou, P., and Triantafillou, E.Towards unbounded machine unlearning.NeurIPS, 2023. Lee et al. (2024) ↑ Lee, C., Chang, J., and Sohn, J.-y.Analysis of using sigmoid loss for contrastive learning.In Dasgupta, S., Mandt, S., and Li, Y. (eds.), Proceedings of The 27th International Conference on Artificial Intelligence and Statistics, volume 238 of Proceedings of Machine Learning Research, pp. 1747–1755. PMLR, 02–04 May 2024.URL https://proceedings.mlr.press/v238/lee24a.html. Li et al. (2024a) ↑ Li, N., Pan, A., Gopal, A., Yue, S., Berrios, D., Gatti, A., Li, J. D., Dombrowski, A.-K., Goel, S., Mukobi, G., et al.The wmdp benchmark: Measuring and reducing malicious use with unlearning.In International Conference on Machine Learning, pp. 28525–28550. PMLR, 2024a. Li et al. (2024b) ↑ Li, W., Li, J., de Witt, C. S., Prabhu, A., and Sanyal, A.Delta-influence: Unlearning poisons via influence functions, 2024b.URL https://arxiv.org/abs/2411.13731. Li et al. (2024c) ↑ Li, X., Zhao, Y., Wu, Z., Zhang, W., Li, R.-H., and Wang, G.Towards effective and general graph unlearning via mutual evolution.In Proceedings of the AAAI Conference on Artificial Intelligence, volume 38, pp. 13682–13690, 2024c. Lingam et al. (2024) ↑ Lingam, V., Akhondzadeh, M. S., and Bojchevski, A.Rethinking label poisoning for GNNs: Pitfalls and attacks.In The Twelfth International Conference on Learning Representations, 2024.URL https://openreview.net/forum?id=J7ioefqDPw. Ma et al. (2022) ↑ Ma, Y., Liu, X., Shah, N., and Tang, J.Is homophily a necessity for graph neural networks?In International Conference on Learning Representations, 2022.URL https://openreview.net/forum?id=ucASPPD9GKN. Maini et al. (2024) ↑ Maini, P., Feng, Z., Schwarzschild, A., Lipton, Z. C., and Kolter, J. Z.TOFU: A task of fictitious unlearning for LLMs.In First Conference on Language Modeling, 2024.URL https://openreview.net/forum?id=B41hNBoWLo. Mao et al. (2024) ↑ Mao, H., Chen, Z., Tang, W., Zhao, J., Ma, Y., Zhao, T., Shah, N., Galkin, M., and Tang, J.Position: Graph foundation models are already here.In Forty-first International Conference on Machine Learning, 2024. Northcutt et al. (2021) ↑ Northcutt, C. G., Athalye, A., and Mueller, J.Pervasive label errors in test sets destabilize machine learning benchmarks.In NeurIPS, 2021. Ouyang et al. (2022) ↑ Ouyang, L., Wu, J., Jiang, X., Almeida, D., Wainwright, C., Mishkin, P., Zhang, C., Agarwal, S., Slama, K., Ray, A., et al.Training language models to follow instructions with human feedback.Advances in neural information processing systems, 35:27730–27744, 2022. Paleka & Sanyal (2023) ↑ Paleka, D. and Sanyal, A.A law of adversarial risk, interpolation, and label noise.In ICLR, 2023. Pawelczyk et al. (2024) ↑ Pawelczyk, M., Di, J. Z., Lu, Y., Kamath, G., Sekhari, A., and Neel, S.Machine unlearning fails to remove data poisoning attacks.arXiv preprint arXiv:2406.17216, 2024. Said et al. (2023) ↑ Said, A., Derr, T., Shabbir, M., Abbas, W., and Koutsoukos, X.A survey of graph unlearning.arXiv preprint arXiv:2310.02164, 2023. Sanyal et al. (2021) ↑ Sanyal, A., Dokania, P. K., Kanade, V., and Torr, P.How benign is benign overfitting?In ICLR, 2021. Sanyal et al. (2022) ↑ Sanyal, A., Hu, Y., and Yang, F.How unfair is private learning?In Uncertainty in Artificial Intelligence, 2022. Schoepf et al. (2024) ↑ Schoepf, S., Foster, J., and Brintrup, A.Potion: Towards poison unlearning.Journal of Data-centric Machine Learning Research, 2024.URL https://openreview.net/forum?id=4eSiRnWWaF. Sun et al. (2019) ↑ Sun, Y., Wang, S., Tang, X., Hsieh, T.-Y., and Honavar, V.Node injection attacks on graphs via reinforcement learning.arXiv preprint arXiv:1909.06543, 2019. Tan et al. (2024) ↑ Tan, J., Sun, F., Qiu, R., Su, D., and Shen, H.Unlink to unlearn: Simplifying edge unlearning in gnns.In Companion Proceedings of the ACM on Web Conference 2024, pp. 489–492, 2024. Veličković et al. (2018) ↑ Veličković, P., Cucurull, G., Casanova, A., Romero, A., Liò, P., and Bengio, Y.Graph attention networks.In International Conference on Learning Representations, 2018. Wang et al. (2020) ↑ Wang, K., Shen, Z., Huang, C., Wu, C.-H., Dong, Y., and Kanakia, A.Microsoft academic graph: When experts are not enough.Quantitative Science Studies, 1(1):396–413, 2020. Warnecke et al. (2023) ↑ Warnecke, A., Pirch, L., Wressnegger, C., and Rieck, K.Machine unlearning of features and labels.In Proc. of the 30th Network and Distributed System Security (NDSS), 2023. Wu & Davison (2005) ↑ Wu, B. and Davison, B. D.Identifying link farm spam pages.In Special Interest Tracks and Posters of the 14th International Conference on World Wide Web, WWW ’05, pp. 820–829, New York, NY, USA, 2005. Association for Computing Machinery.ISBN 1595930515.doi: 10.1145/1062745.1062762.URL https://doi.org/10.1145/1062745.1062762. Wu et al. (2023a) ↑ Wu, J., Yang, Y., Qian, Y., Sui, Y., Wang, X., and He, X.Gif: A general graph unlearning strategy via influence function.In Proceedings of the ACM Web Conference 2023, pp. 651–661, 2023a. Wu et al. (2023b) ↑ Wu, K., Shen, J., Ning, Y., Wang, T., and Wang, W. H.Certified edge unlearning for graph neural networks.In Proceedings of the 29th ACM SIGKDD Conference on Knowledge Discovery and Data Mining, pp. 2606–2617, 2023b. Wu et al. (2022) ↑ Wu, S., Sun, F., Zhang, W., Xie, X., and Cui, B.Graph neural networks in recommender systems: A survey.ACM Comput. Surv., 55(5), December 2022.ISSN 0360-0300.doi: 10.1145/3535101.URL https://doi.org/10.1145/3535101. Xi et al. (2023) ↑ Xi, Z., Du, T., Li, C., Pang, R., Ji, S., Luo, X., Xiao, X., Ma, F., and Wang, T.On the security risks of knowledge graph reasoning.In Proceedings of the 32nd USENIX Conference on Security Symposium, SEC ’23, USA, 2023. USENIX Association.ISBN 978-1-939133-37-3. Yao et al. (2024) ↑ Yao, Y., Xu, X., and Liu, Y.Large language model unlearning.Advances in Neural Information Processing Systems, 37:105425–105475, 2024. Yuan et al. (2024) ↑ Yuan, X., Zhang, C., Tian, Y., Ye, Y., and Zhang, C.Mitigating emergent robustness degradation while scaling graph learning.In The Twelfth International Conference on Learning Representations, 2024.URL https://openreview.net/forum?id=Koh0i2u8qX. Zhang et al. (2023) ↑ Zhang, C., Tian, Y., Ju, M., Liu, Z., Ye, Y., Chawla, N., and Zhang, C.Chasing all-round graph representation robustness: Model, training, and optimization.In The Eleventh International Conference on Learning Representations, 2023.URL https://openreview.net/forum?id=7jk5gWjC18M. Zhang et al. (2019) ↑ Zhang, H., Zheng, T., Gao, J., Miao, C., Su, L., Li, Y., and Ren, K.Data poisoning attack against knowledge graph embedding.In Proceedings of the Twenty-Eighth International Joint Conference on Artificial Intelligence, IJCAI-19, pp. 4853–4859. International Joint Conferences on Artificial Intelligence Organization, 7 2019.doi: 10.24963/ijcai.2019/674.URL https://doi.org/10.24963/ijcai.2019/674. Zhang et al. (2022) ↑ Zhang, Z., Chen, L., Zhong, F., Wang, D., Jiang, J., Zhang, S., Jiang, H., Zheng, M., and Li, X.Graph neural network approaches for drug-target interactions.Current Opinion in Structural Biology, 73:102327, 2022.ISSN 0959-440X.doi: https://doi.org/10.1016/j.sbi.2021.102327.URL https://www.sciencedirect.com/science/article/pii/S0959440X2100169X. Zhao et al. (2024) ↑ Zhao, T., Chen, J., Ru, Y., Lin, Q., Geng, Y., and Liu, J.Untargeted adversarial attack on knowledge graph embeddings.In Proceedings of the 47th International ACM SIGIR Conference on Research and Development in Information Retrieval, pp. 1701–1711, 2024. Zügner & Günnemann (2019) ↑ Zügner, D. and Günnemann, S.Adversarial attacks on graph neural networks via meta learning.In International Conference on Learning Representations (ICLR), 2019. Zügner et al. (2018) ↑ Zügner, D., Akbarnejad, A., and Günnemann, S.Adversarial attacks on neural networks for graph data.In Proceedings of the 24th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining, KDD ’18, pp. 2847–2856, New York, NY, USA, 2018. Association for Computing Machinery.ISBN 9781450355520.doi: 10.1145/3219819.3220078.URL https://doi.org/10.1145/3219819.3220078. Appendix
A. Proof of Theoretical Results A.1. Proof of Receptive Field TheoremA.1 A.2. Proof of Convergence Theorem A.2 A.3. Proof of IC Attack Theorem A.3
B. Data Manipulation StatisticsB
C. Formal Description of Cognac1
D. Results Showing The Breadth of Applicability D.1. Results on GATD.1 D.2. Performance on large 𝑆 𝑓 D.2 D.3. Results on a trigger poisoning attackD.3
E. Convergence and Ablations E.1. Convergence AnalysisE.1 E.2. Analysis of Affected Neighbors MethodE.2 E.3. Comparing our sampling method to MEGU’s7 E.4. Analysis of Acc aff ReductionE.4
F. Detailed Experimental Setup F.1. Hyperparameter TuningF.1 F.2. Unlearning TimesF.2
Appendix AProof of Theoretical Results A.1Proof of Lemma 3.2 Lemma.
Let 𝐺
( 𝑉 , 𝐸 ) be an undirected graph, and let 𝒩 ( 𝑣 ) denote the 1-hop neighborhood of node 𝑣 . For a node 𝑠 ∈ 𝑉 , let 𝒩 𝑛 ( 𝑠 ) denote the 𝑛 -hop neighborhood of 𝑠 , defined recursively as:
𝒩 𝑛 ( 𝑠 )
{
{
𝑠
}
if
𝑛
0 ,
⋃ 𝑣 ∈ 𝒩 𝑛 − 1 ( 𝑠 ) 𝒩 ( 𝑣 )
if 𝑛
0
In an 𝑛 -layer GNN, the representation 𝑧 𝑠 of node 𝑠 can affect the representations 𝑧 𝑣 of nodes 𝑣 only if 𝑣 ∈ 𝒩 𝑛 ( 𝑠 ) . For any 𝑣 ∉ 𝒩 𝑛 ( 𝑠 ) , 𝑧 𝑣 is independent of 𝑧 𝑠 .
Proof.
We prove the claim by induction on the number of layers 𝑙 in the GNN.
Base Case ( 𝑙
0 ): At layer 𝑙
0 , the representation of a node 𝑣 , ℎ 𝑣 ( 0 ) , is initialized based only on the node’s features. Thus, 𝑧 𝑠
ℎ 𝑠 ( 0 ) can only affect ℎ 𝑠 ( 0 ) itself, and no other node 𝑣 ≠ 𝑠 is affected.
Inductive Hypothesis: Assume that at layer 𝑙 , the representation 𝑧 𝑠
ℎ 𝑠 ( 𝑙 ) affects only the representations ℎ 𝑣 ( 𝑙 ) of nodes 𝑣 ∈ 𝒩 𝑙 ( 𝑠 ) , the 𝑙 -hop neighborhood of 𝑠 .
Inductive Step: At layer 𝑙 + 1 , the representation ℎ 𝑣 ( 𝑙 + 1 ) of a node 𝑣 is computed as:
ℎ 𝑣 ( 𝑙 + 1 )
𝜙 ( ℎ 𝑣 ( 𝑙 ) , 𝜓 ( { ℎ 𝑢 ( 𝑙 ) : 𝑢 ∈ 𝒩 ( 𝑣 ) } ) )
By the inductive hypothesis, ℎ 𝑢 ( 𝑙 ) depends only on nodes in 𝒩 𝑙 ( 𝑢 ) . Therefore, ℎ 𝑣 ( 𝑙 + 1 ) depends on nodes in:
⋃ 𝑢 ∈ 𝒩 ( 𝑣 ) 𝒩 𝑙 ( 𝑢 )
Since 𝒩 𝑙 + 1 ( 𝑣 )
⋃ 𝑢 ∈ 𝒩 ( 𝑣 ) 𝒩 𝑙 ( 𝑢 ) , ℎ 𝑣 ( 𝑙 + 1 ) depends only on nodes in 𝒩 𝑙 + 1 ( 𝑣 ) .
For 𝑣 to be influenced by 𝑧 𝑠 , there must exist a path of length at most 𝑙 + 1 from 𝑠 to 𝑣 , i.e., 𝑣 ∈ 𝒩 𝑙 + 1 ( 𝑠 ) .
Conclusion: By induction, after 𝑛 layers, 𝑧 𝑠 can only affect nodes in 𝒩 𝑛 ( 𝑠 ) . For any 𝑣 ∉ 𝒩 𝑛 ( 𝑠 ) , the representation 𝑧 𝑣 is independent of 𝑧 𝑠 . ∎
A.2Proof of Theorem 3.3 Theorem.
Let 𝜃 ∈ ℝ 𝑑 parameterize a GNN generating embeddings 𝑧 𝑣 , 𝑧 𝑝 , 𝑧 𝑛 ∈ ℝ 𝑘 for triplets ( 𝑣 , 𝑝 , 𝑛 ) . Let ℒ ( 𝜃 )
− 𝔼 [ log 𝜎 ( 𝑧 𝑣 ⊤ 𝑧 𝑝 ) + log 𝜎 ( − 𝑧 𝑣 ⊤ 𝑧 𝑛 ) ] and 𝑆 ( 𝜃 )
𝔼 [ 𝑧 𝑣 ⊤ 𝑧 𝑝 − 𝑧 𝑣 ⊤ 𝑧 𝑛 ] .
Then under the assumptions stated below, gradient descent on ℒ ( 𝜃 ) with step size 𝜂 ≤ 1 𝐿 (where 𝐿 is the Lipschitz constant of ∇ 𝜃 ℒ ) guarantees:
𝑆 ( 𝜃 𝑡 + 1 ) ≥ 𝑆 ( 𝜃 𝑡 ) ∀ 𝑡 ≥ 0 ,
and at convergence, 𝑆 ( 𝜃 ∗ )
𝑆 ( 𝜃 0 ) .
Assumption A.1 (Differentiability).
𝑧 𝑣 ⊤ 𝑧 𝑝 and 𝑧 𝑣 ⊤ 𝑧 𝑛 are differentiable in 𝜃 .
Assumption A.2 (Convexity).
ℒ ( 𝜃 ) is convex in 𝜃 .
Assumption A.3 (Bounded Gradients).
‖ ∇ 𝜃 ( 𝑧 𝑣 ⊤ 𝑧 𝑝 ) ‖ ≤ 𝐺 and ‖ ∇ 𝜃 ( 𝑧 𝑣 ⊤ 𝑧 𝑛 ) ‖ ≤ 𝐺 for some 𝐺
0 .
Proof.
Let 𝑎
𝑧 𝑣 ⊤ 𝑧 𝑝 and 𝑏
𝑧 𝑣 ⊤ 𝑧 𝑛 . Then:
ℒ ( 𝜃 )
− 𝔼 [ log 𝜎 ( 𝑎 ) + log 𝜎 ( − 𝑏 ) ]
Under Assumption 1 and using 𝑑 𝑑 𝑥 ( log 𝜎 ( 𝑥 ) )
1 − 𝜎 ( 𝑥 ) and 𝑑 𝑑 𝑥 ( log 𝜎 ( − 𝑥 ) )
− 𝜎 ( 𝑥 ) , we compute:
∇ 𝜃 ℒ
− 𝔼 [ ( 1 − 𝜎 ( 𝑎 ) ) ⏟ Positive ∇ 𝜃 𝑎 − 𝜎 ( 𝑏 ) ⏟ Positive ∇ 𝜃 𝑏 ]
The gradient of 𝑆 ( 𝜃 ) is:
∇ 𝜃 𝑆
𝔼 [ ∇ 𝜃 𝑎 − ∇ 𝜃 𝑏 ]
The gradient of ℒ can be rewritten as:
∇ 𝜃 ℒ
− 𝔼 [ 𝜎 ( − 𝑎 ) ∇ 𝜃 𝑎 − 𝜎 ( 𝑏 ) ∇ 𝜃 𝑏 ]
Both 𝜎 ( − 𝑎 ) and 𝜎 ( 𝑏 ) are positive for all finite 𝑎 , 𝑏 . Thus, ∇ 𝜃 ℒ is a negatively weighted combination of ∇ 𝜃 𝑎 and ∇ 𝜃 𝑏 , while ∇ 𝜃 𝑆 is their unweighted difference.
Next, we show the monotonic improvement of 𝑆 ( 𝜃 ) . Consider a gradient descent update:
𝜃 𝑡 + 1
𝜃 𝑡 − 𝜂 ∇ 𝜃 ℒ
The change in 𝑆 ( 𝜃 ) is:
Δ 𝑆
𝑆 ( 𝜃 𝑡 + 1 ) − 𝑆 ( 𝜃 𝑡 )
To simplify the analysis, we’ll use a first-order approximation. Specifically, for small updates, we approximate:
𝑆 ( 𝜃 𝑡 + 1 ) ≈ 𝑆 ( 𝜃 𝑡 ) + ⟨ ∇ 𝜃 𝑆 , 𝜃 𝑡 + 1 − 𝜃 𝑡 ⟩ ,
where ⟨ ⋅ , ⋅ ⟩ denotes the inner product. Substituting 𝜃 𝑡 + 1
𝜃 𝑡 − 𝜂 ∇ 𝜃 ℒ ( 𝜃 𝑡 ) , we get:
Δ 𝑆 ≈ − 𝜂 ⟨ ∇ 𝜃 𝑆 , ∇ 𝜃 ℒ ⟩
Substituting ∇ 𝜃 𝑆 and ∇ 𝜃 ℒ :
⟨ ∇ 𝜃 𝑆 , ∇ 𝜃 ℒ ⟩
𝔼 [ ( ∇ 𝜃 𝑎 − ∇ 𝜃 𝑏 ) ⊤ ( ( 1 − 𝜎 ( 𝑎 ) ) ∇ 𝜃 𝑎 − 𝜎 ( 𝑏 ) ∇ 𝜃 𝑏 ) ]
Δ 𝑆 ≈ 𝜂 𝔼 [ 𝜎 ( − 𝑎 ) ‖ ∇ 𝜃 𝑎 ‖ 2 + 𝜎 ( 𝑏 ) ‖ ∇ 𝜃 𝑏 ‖ 2 − 𝜎 ( − 𝑎 ) ∇ 𝜃 𝑏 ⊤ ∇ 𝜃 𝑎 − 𝜎 ( 𝑏 ) ∇ 𝜃 𝑎 ⊤ ∇ 𝜃 𝑏 ]
But since the dot product is commutative for real-valued column vectors, i.e., ∇ 𝜃 𝑏 ⊤ ∇ 𝜃 𝑎
∇ 𝜃 𝑎 ⊤ ∇ 𝜃 𝑏 ,
Δ 𝑆 ≈ 𝜂 𝔼 [ 𝜎 ( − 𝑎 ) ‖ ∇ 𝜃 𝑎 ‖ 2 + 𝜎 ( 𝑏 ) ‖ ∇ 𝜃 𝑏 ‖ 2 − ( 𝜎 ( − 𝑎 ) + 𝜎 ( 𝑏 ) ) ( ∇ 𝜃 𝑎 ⊤ ∇ 𝜃 𝑏 ) ]
This expression contains both positive and negative terms. However, we now use the fact that the sigmoid function 𝜎 ( 𝑥 ) satisfies 0 < 𝜎 ( 𝑥 ) < 1 for all real values of 𝑥 , which means:
( ( 1 − 𝜎 ( 𝑎 ) )
𝜎 ( − 𝑎 )
0 ) for ( 𝑎 ∈ ℝ ) and ( 𝜎 ( 𝑏 )
0 ) for ( 𝑏 ∈ ℝ )
Thus, both terms 𝜎 ( − 𝑎 ) ‖ ∇ 𝜃 𝑎 ‖ 2 and 𝜎 ( 𝑏 ) ‖ ∇ 𝜃 𝑏 ‖ 2 are positive. On the other hand, the cross-product terms ∇ 𝜃 𝑎 ⊤ ∇ 𝜃 𝑏 might be negative, but their magnitude is bounded by the gradients ‖ ∇ 𝜃 𝑎 ‖ and ‖ ∇ 𝜃 𝑏 ‖ , which are constrained by the assumption of bounded gradients (Assumption 3). Under this assumption, the gradients ‖ ∇ 𝜃 𝑎 ‖ and ‖ ∇ 𝜃 𝑏 ‖ are bounded by 𝐺 , i.e., ‖ ∇ 𝜃 𝑎 ‖ ≤ 𝐺 and ‖ ∇ 𝜃 𝑏 ‖ ≤ 𝐺 . Therefore, the cross-product terms are also bounded:
| ∇ 𝜃 𝑎 ⊤ ∇ 𝜃 𝑏 | ≤ 𝐺 2
Now we can write the change in 𝑆 ( 𝜃 ) as:
Δ 𝑆 ≈ 𝜂 𝔼 [ 𝜎 ( − 𝑎 ) ‖ ∇ 𝜃 𝑎 ‖ 2 + 𝜎 ( 𝑏 ) ‖ ∇ 𝜃 𝑏 ‖ 2 − 2 𝐺 2 ]
For sufficiently small 𝜂 , the positive terms 𝜎 ( − 𝑎 ) ‖ ∇ 𝜃 𝑎 ‖ 2 and 𝜎 ( 𝑏 ) ‖ ∇ 𝜃 𝑏 ‖ 2 will dominate the cross-product terms, ensuring that Δ 𝑆 ≥ 0 .
Δ 𝑆 ≥ 𝜂 𝔼 [ 𝜎 ( − 𝑎 ) ‖ ∇ 𝜃 𝑎 ‖ 2 + 𝜎 ( 𝑏 ) ‖ ∇ 𝜃 𝑏 ‖ 2 − 2 𝐺 2 ]
For sufficiently small 𝜂 , Δ 𝑆 ≥ 0 . Thus, 𝑆 ( 𝜃 𝑡 + 1 ) ≥ 𝑆 ( 𝜃 𝑡 ) .
By convexity (Assumption 2), gradient descent converges to a global minimum 𝜃 ∗ where ∇ 𝜃 ℒ ( 𝜃 ∗ )
0 . At 𝜃 ∗ :
𝔼 [ 𝜎 ( − 𝑎 ) ∇ 𝜃 𝑎 ]
𝔼 [ 𝜎 ( 𝑏 ) ∇ 𝜃 𝑏 ]
This equality holds only if 𝜎 ( − 𝑎 ) → 0 and 𝜎 ( 𝑏 ) → 0 , which occurs when 𝑎 → ∞ and 𝑏 → − ∞ . Thus:
𝑆 ( 𝜃 ∗ )
𝔼 [ 𝑧 𝑣 ⊤ 𝑧 𝑝 − 𝑧 𝑣 ⊤ 𝑧 𝑛 ] ≥ 𝑆 ( 𝜃 0 )
Strict inequality 𝑆 ( 𝜃 ∗ )
𝑆 ( 𝜃 0 ) follows from the monotonic improvement at each step.
∎
A.3Proof of Theorem 3.1 Theorem.
Let 𝐺
( 𝑉 , 𝐸 , 𝑋 ) be a graph with node set 𝑉 , edge set 𝐸 , and features 𝑋 . Let 𝐶 1 , 𝐶 2 ⊂ 𝑉 be two distinct classes with ground-truth labels 𝑦 𝑖 ∈ { 𝐶 1 , 𝐶 2 } . Suppose an Interclass Confusion (IC) attack is applied as follows:
1.
Select subsets 𝑆 ′ 𝐶 1 ⊂ 𝐶 1 and 𝑆 ′ 𝐶 2 ⊂ 𝐶 2 , each of size 𝑛 2 , forming 𝑆 ′
𝑆 ′ 𝐶 1 ∪ 𝑆 ′ 𝐶 2
2.
Swap labels of 𝛼
fraction of 𝑆 ′ , creating a confused set 𝑆 𝑓
3.
Train a GNN model 𝑀 on ( 𝑆 ∖ 𝑆 ′ ) ∪ 𝑆 𝑓
Let 𝜙 𝑀 ( 𝐶 1 ) , 𝜙 𝑀 ( 𝐶 2 ) ∈ ℝ 𝑑 denote the mean embeddings of 𝐶 1 and 𝐶 2 , and 𝒟 ( 𝜙 𝑀 ( 𝐶 1 ) , 𝜙 𝑀 ( 𝐶 2 ) ) be the Wasserstein-2 distance between their embedding distributions. Assume:
Assumption A.1 (Homophily preserving GNNs).
The GNN uses 𝐿 -layers of homophily-preserving message passing (e.g., mean aggregation).
Assumption A.2 (Homophily).
The graph exhibits 𝜂 -homophily, where intra-class edge density dominates inter-class.
Assumption A.3 (Cross-Entropy Loss).
The loss function ℒ includes cross-entropy and graph smoothness regularization.
Then, there exists a degradation term Δ
0 such that:
𝔼 [ 𝒟 ( 𝜙 𝑀 ( 𝐶 1 ) , 𝜙 𝑀 ( 𝐶 2 ) ) ] ≤ 𝔼 [ 𝒟 ( 𝜙 𝑀 clean ( 𝐶 1 ) , 𝜙 𝑀 clean ( 𝐶 2 ) ) ] − Δ ,
where Δ ∝ 𝛼 ⋅ 𝜂 ⋅ ( 1 − 1 𝐿 ) .
Proof.
Step 1: Embedding Process Formalization
The GNN computes node embeddings via 𝐿 -layer message passing. For node 𝑣 , the embedding 𝜙 ( 𝑙 ) ( 𝑣 ) at layer 𝑙 is:
𝜙 ( 𝑙 ) ( 𝑣 )
𝜎 ( 𝐖 ( 𝑙 ) ⋅ AGG ( { 𝜙 ( 𝑙 − 1 ) ( 𝑢 ) ∣ 𝑢 ∈ 𝒩 ( 𝑣 ) } ) ) ,
where AGG is a mean aggregator, 𝐖 ( 𝑙 ) are learnable weights, and 𝜎 is a nonlinearity. The final embedding 𝜙 𝑀 ( 𝑣 )
𝜙 ( 𝐿 ) ( 𝑣 ) .
Step 2: Loss Function and Label Noise
The training loss ℒ
ℒ CE + 𝜆 ℒ reg , where:
•
ℒ CE
− 1 | 𝑆 | ∑ 𝑣 ∈ 𝑆 𝑦 𝑣 log 𝑦 ^ 𝑣 (cross-entropy)
•
ℒ reg
1 | 𝐸 | ∑ ( 𝑢 , 𝑣 ) ∈ 𝐸 ‖ 𝜙 𝑀 ( 𝑢 ) − 𝜙 𝑀 ( 𝑣 ) ‖ 2 (smoothness regularization)
For 𝑆 𝑓 , labels 𝑦 𝑣 are swapped between 𝐶 1 and 𝐶 2 . Let ℰ noise
{ 𝑣 ∈ 𝑆 𝑓 ∣ 𝑦 𝑣 is incorrect } . The corrupted ℒ CE forces conflicting gradients for 𝑣 ∈ ℰ noise , pulling 𝜙 𝑀 ( 𝑣 ) toward the wrong class centroid.
Step 3: Bias in Mean Embeddings
Let 𝜇 𝐶 1 , 𝜇 𝐶 2 be the mean embeddings of 𝐶 1 , 𝐶 2 under 𝑀 clean . After IC attack, for 𝑣 ∈ ℰ noise :
𝔼 [ 𝜙 𝑀 ( 𝑣 ) ]
𝛼 𝜇 𝐶 2 + ( 1 − 𝛼 ) 𝜇 𝐶 1 ( if 𝑣 ∈ 𝐶 1 )
By linearity of expectation, the perturbed mean for 𝐶 1 becomes:
𝜇 𝐶 1 ′
𝜇 𝐶 1 − 𝛼 ( 𝜇 𝐶 1 − 𝜇 𝐶 2 )
Similarly, 𝜇 𝐶 2 ′
𝜇 𝐶 2 + 𝛼 ( 𝜇 𝐶 1 − 𝜇 𝐶 2 ) . Thus, the distance between means reduces by 2 𝛼 ‖ 𝜇 𝐶 1 − 𝜇 𝐶 2 ‖ .
Step 4: Message-Passing Amplification
Under 𝜂 -homophily (where 𝜂 is the ‘extent’ of homophily), neighbors of 𝑣 ∈ ℰ noise are likely in 𝐶 1 . The smoothness term ℒ reg propagates the corrupted embedding of 𝑣 to its neighbors, perturbing their embeddings. After 𝐿 layers, the influence of ℰ noise spreads to ∼ 𝜂 𝐿 | 𝑉 | nodes. This amplifies the mean embedding shift by a factor 𝜂 ( 1 − 1 𝐿 ) .
Step 5: Wasserstein Distance Bound
The Wasserstein-2 distance between 𝜙 𝑀 ( 𝐶 1 ) and 𝜙 𝑀 ( 𝐶 2 ) is dominated by the mean shift and covariance distortion. Using the Fréchet inequality:
𝒟 2 ≤ ‖ 𝜇 𝐶 1 ′ − 𝜇 𝐶 2 ′ ‖ 2 + Tr ( Σ 𝐶 1 + Σ 𝐶 2 − 2 ( Σ 𝐶 1 Σ 𝐶 2 ) 1 / 2 )
From Steps 3 and 4, ‖ 𝜇 𝐶 1 ′ − 𝜇 𝐶 2 ′ ‖ 2
( 1 − 2 𝛼 𝜂 ( 1 − 1 𝐿 ) ) ‖ 𝜇 𝐶 1 − 𝜇 𝐶 2 ‖ 2 . Thus,
𝔼 [ 𝒟 ( 𝜙 𝑀 ( 𝐶 1 ) , 𝜙 𝑀 ( 𝐶 2 ) ) ] ≤ 𝔼 [ 𝒟 ( 𝜙 𝑀 clean ( 𝐶 1 ) , 𝜙 𝑀 clean ( 𝐶 2 ) ) ] − Δ ,
where Δ
𝛼 𝜂 ( 1 − 1 𝐿 ) ‖ 𝜇 𝐶 1 − 𝜇 𝐶 2 ‖ 2 .
Conclusion
The IC attack reduces the separability of 𝐶 1 and 𝐶 2 in the embedding space by biasing their mean embeddings toward each other and amplifying this bias via graph convolutions. The degradation Δ scales with 𝛼 , 𝜂 , and the depth 𝐿 , confirming representation entanglement. ∎
Appendix BData Manipulation Statistics
Table 4 presents the manipulation statistics for all datasets used in our study. We introduce a significant variation in the degree of manipulation to achieve two key objectives: (1) in some datasets, a lower percentage of manipulated nodes or edges is sufficient to induce noticeable performance degradation, while others require more extensive modifications, and (2) we aim to evaluate unlearning performance across a broad spectrum of manipulation scales. For instance, PubMed undergoes the highest level of manipulation, with nearly 39% of training nodes affected and 33.84% additional edges introduced, whereas datasets like CS and CoraFull experience minimal modifications. Additionally, for OGB-Arxiv, a significantly larger dataset, only 3.14% of training nodes are manipulated. This diverse manipulation strategy ensures a rigorous and comprehensive evaluation of unlearning effectiveness across different graph structures.
Table 4:Dataset manipulation statistics. Statistics of manipulations across datasets. The percentage of nodes and edges added are relative to the existing number of training nodes and the number of edges in the graph, respectively. Dataset Training Nodes Nodes Manipulated ( % ) Edges Added ( % )
CoraFull 11 , 280
1.45
1.20
Cora 1 , 491
14.88
17.26
CiteSeer 1 , 272
14.78
20.38
DBLP 9 , 714
10.50
5.78
PubMed 11 , 830
38.96
33.84
OGB-arXiv 90 , 941
3.14 - CS 10 , 999
2.25
1.83
Physics 20 , 695
8.17
5.04
Photos 4 , 492
11.40
5.04
Computers 8 , 028
3.14
0.814 Appendix CFormal Description of Cognac
In this section, we outline the procedure of our proposed unlearning method, Cognac, designed to effectively remove the influence of manipulated data from GNNs. First, the algorithm identifies the nodes affected by the manipulation, as well as their corresponding positive and negative samples, and then alternates between applying CoGN and AC ↯ DC to unlearn their influence. The key steps include identifying the affected nodes, performing contrastive learning to re-optimize the embeddings, and minimizing classification loss on the unaffected nodes while maximizing it on the discovered manipulated set ( 𝑆 𝑓 ). The complete algorithm is detailed in Algorithm 1.
Algorithm 1 Cognac 0: GNN 𝑀 , Graph 𝐺
( 𝑉 , 𝐸 , 𝑋 ) , Deletion set 𝑆 𝑓 , Hyperparameters Θ 0: Unlearned GNN 𝑀 ∗ 1: 𝑆 ← IdentifyAffectedNodes ( 𝑀 , 𝑋 , 𝑆 𝑓 , 𝐸 , Θ ) 2: 𝑃 ← SamplePositives ( 𝑆 , 𝐸 , 𝑆 𝑓 ) 3: 𝑁 ← SampleNegatives ( 𝑆 , 𝑆 𝑓 ) 4: // Overall unlearning process 5: for outer_epoch = 1 to Θ total_epochs do 6: // Contrastive unlearning phase 7: for epoch = 1 to Θ contrast_epochs do 8: 𝑍 ← 𝑀 ( 𝑋 ) 9: ℒ 𝑐 ← ∑ 𝑣 ∈ 𝑆 ( − log ( 𝜎 ( 𝑍 𝑣 𝑇 𝑍 𝑃 ) ) − log ( 𝜎 ( − 𝑍 𝑣 𝑇 𝑍 𝑁 ) ) ) 10: 𝑀 ← Optimize ( 𝑀 , ℒ 𝑐 ) 11: end for 12: // Gradient ascent on 𝑆 𝑓 , and gradient descent on 𝑉 ∖ 𝑆 𝑓 13: for epoch = 1 to Θ ascent_descent_epochs do 14: ℒ 𝑎 ← − CrossEntropy ( 𝑀 ( 𝑋 ) 𝑆 𝑓 , 𝑌 𝑆 𝑓 ) 15: 𝑀 ← Optimize ( 𝑀 , ℒ 𝑎 ) 16: ℒ 𝑑 ← CrossEntropy ( 𝑀 ( 𝑋 ) 𝑉 ∖ 𝑆 𝑓 , 𝑌 𝑉 ∖ 𝑆 𝑓 ) 17: 𝑀 ← Optimize ( 𝑀 , ℒ 𝑑 ) 18: end for 19: end for 20: return 𝑀 21: 22: function IdentifyAffectedNodes( 𝑀 , 𝑋 , 𝑆 𝑓 , 𝐸 , Θ ) 23: 𝑋 ′ ← InvertFeatures ( 𝑋 , 𝑆 𝑓 , 𝐸 ) 24: Δ ← | 𝑀 ( 𝑋 ′ ) − 𝑀 ( 𝑋 ) | 25: return TopK ( Δ , Θ 𝑘 ) 26: end function
Appendix DResults Showing The Breadth of Applicability of Cognac D.1Results on GAT
To provide a comprehensive comparison between Cognac and other methods, we provide results on commonly used GNN backbone architectures - GCN and GAT.
Graph Convolutional Network (GCN) is a method for semi-supervised classification of graph-structured data. It employs an efficient layer-wise propagation rule derived from a first-order approximation of spectral convolutions on graphs.
Graph Attention Network (GAT) employs computationally efficient masked self-attention layers that assign varying importance to neighborhood nodes without needing the complete graph structure upfront, thereby overcoming many theoretical limitations of earlier spectral-based methods.
Figure 6:Node Unlearning results for Cora with GAT backbone. Comparison of Acc aff for the unlearning methods on GAT trained on Cora for different values of ( 𝑆 𝑓 / 𝑆 𝑚 ) . We see that Cognac outperforms all baselines.
Figure 6 shows that Cognac also performs competitively with a GAT backbone. When 5% of 𝑆 𝑚 is known, SCRUB performs similarly to Cognac, within the standard deviation. For higher fractions, we achieve greater Acc aff than the benchmark graph unlearning methods with large margins, often beating the performance of retraining the GNN from scratch. These results indicate that benchmark graph unlearning methods used for comparison cannot recover from the impact of the label flip poison. In contrast, our method is much closer to Oracle’s performance.
D.2Additional Experiments on Large 𝑆 𝑓
To stress-test Cognac’s performance - as methods could potentially degrade as the size of 𝑆 𝑓 grows - we conduct an experiment where we choose a significant fraction of the training nodes of the Amazon, DBLP, and Physics datasets, to be attacked (by the binary label flip attack) and marked for deletion. The method performs competitively even at this large deletion size. Table 5 demonstrates these results.
Table 5:Evaluation on larger sizes of 𝑆 𝑓 . Cognac performs well across datasets (Amazon, DBLP, Physics), even when a significant fraction of the total training nodes are present in 𝑆 𝑓 . Method Amazon ( 25 % ) DBLP ( 29.4 % ) Physics ( 14.8 % )
Acc rem
Acc aff
Acc rem
Acc aff
Acc rem
Acc aff
Oracle 92.9
95.6
72.9
86.0
95.2
95.1
Original 92.5
49.0
74.9
57.9
58.9
95.4
Cognac 92.4
83.7
82.3
81.7
90.7
95.0
GNNDelete 27.7
49.7
45.0
49.6
1.4
37.3
SCRUB 92.3
72.6
77.5
82.1
77.9
94.9 D.3Results on a trigger poisoning attack
We expand the analysis to a wider range of attack scenarios, now covering all major poisoning types (label, graph structure, and feature). Our feature attack injects trigger a pattern into the feature vectors of select nodes, and assigning a fixed spurious label, and reduces accuracy on the target distribution. Despite not being the strongest possible attack, our implementation provides sufficient signal for evaluation - most unlearning methods struggle, while Cognac matches and even outperforms retraining performance. Complete results are provided in Table 6.
The attacker selects a subset of victim nodes 𝑆 𝑝 ⊂ 𝑉 . For each node 𝑣 ∈ 𝑆 𝑝 , its original feature vector 𝑥 𝑣 is modified to 𝑥 𝑣 ′ by setting specific trigger feature indices 𝑗 ∈ 𝐼 𝑡 to 1 (i.e., 𝑥 𝑣 , 𝑗 ′
1 ∀ 𝑗 ∈ 𝐼 𝑡 , while other features 𝑥 𝑣 , 𝑗 for 𝑗 ∉ 𝐼 𝑡 remain unchanged). Subsequently, all nodes 𝑣 ∈ 𝑆 𝑝 are assigned a fixed target label 𝑦 𝑡 𝑎 𝑟 𝑔 𝑒 𝑡 . The choice of victim nodes here is random within the victim class, and the number is chosen to maintain stealth while maximising Attack Success Rate (the percentage of manipulated nodes in the test set that are successfully classified as the target class).
This use of a localized feature pattern as a trigger is analogous to patch-based backdoor attacks in image poisoning, such as those introduced by BadNets (Gu et al., 2017).
Table 6:Results for unlearning the trigger-based feature manipulation. Cognac performs strongly across datasets, maintaining a high Acc aff and Acc rem . Retrain fails to recover accuracy on the victim class for the Cora dataset, while SCRUB fails to do so for both Cora and CS. GNNDelete and MEGU, the graph unlearning baselines, fail to remove the poison. Method Photos Cora CS
Acc rem
Acc aff
Acc rem
Acc aff
Acc rem
Acc aff
Original 95.0 ± 0.0
33.9 ± 0.0
67.4 ± 0.0
25.6 ± 0.0
89.3 ± 0.0
0.0 ± 0.0
retrain 94.1 ± 0.5
92.0 ± 1.7
68.3 ± 0.4
46.9 ± 6.8
93.0 ± 0.2
92.8 ± 2.3
gnndelete 17.8 ± 0.0
0.0 ± 0.0
65.7 ± 0.2
40.8 ± 9.1
68.0 ± 0.0
0.0 ± 0.0
megu 81.5 ± 8.7
17.5 ± 23.1
61.5 ± 0.3
1.0 ± 1.0
88.7 ± 0.1
0.0 ± 0.0
scrub 93.8 ± 0.0
92.7 ± 0.0
68.0 ± 0.0
64.1 ± 0.0
92.7 ± 0.1
72.1 ± 21.8
Cognac 94.6 ± 0.0
91.1 ± 0.0
67.3 ± 0.0
78.2 ± 0.0
93.3 ± 0.0
93.6 ± 0.6 Appendix EConvergence and Ablations of Cognac E.1Convergence
We now discuss the convergence properties of Cognac. Plots in Figure 7 describe the losses of each of the components of our method (contrastive, ascent, descent) after the last epoch of every step, the meaning of which should be clear from Algorithm 1: Line 4 (which we denote as num_steps). The loss plots are constructed over the best hyperparameters, and we would likely not see such convergence trends with sub-optimal hyperparameters, which may provide insights to improve performance when it’s used in other settings as well.
Figure 7:Convergence of the losses across unlearning steps. The ascent loss on 𝑆 𝑚 continually increases as expected, the descent loss on 𝑆 ∖ 𝑆 𝑚 converges, and the contrastive loss exhibits a low plateau after an initial overshoot, implying it may have learned discriminative features. E.2Analysis of method used to find affected neighbors
Our strategy to find affected neighbors is likely not perfect for finding the most affected nodes, and more sophisticated influence functions, such as the one presented in (Chen et al., 2023), could be used to potentially improve performance. Still, we note that it achieves a 5 % higher Acc aff than while choosing random 𝑘 % nodes in the 𝑛 -hop neighborhood (where 𝑛 is the number of layers of message passing) while being cheap to compute: we only require a single forward pass over the model with the inverted features. Interestingly, Figure 8 also shows that even if the GNN is not well-trained, if we choose the top 𝑘 % affected nodes, the unlearning performance does not change much, while still being noticeably better than when we use a random 𝑘 % of the neighbors.
Figure 9 (left) shows that there are no noticeable changes in taking a smaller or larger 𝑘 % . However, removing this step entirely ( 𝑘
0 % ) results in worse performance, suggesting that performing contrastive unlearning on even a small 𝑘 % is significant. Additionally, by keeping this percentage small, we ensure computational efficiency without diminishing performance (Figure 9 (right)), which is essential for unlearning methods.
Figure 8:Effect of how well-trained the GNN is on top- 𝑘 % affected neighbor identification. Here 𝑘
4 . The x-axis represents the epoch at which we used GNN representations to identify the most affected neighbors for Cognac. The y-axis reports the unlearning performance after contrastive unlearning on these identified nodes using the final model. The red line contains performance after picking a random subset from the 𝑛 -hop neighborhood. Affected neighborhood identification using top- 𝑘 % logit change is more effective even with an extremely undertrained GNN. Figure 9:Effect of 𝑘 on unlearning ( Acc aff ), utility ( Acc rem ) and efficiency when identifying top- 𝑘 % affected nodes for contrastive unlearning. (Left) Cognac effectiveness sharply improves beyond 𝑘
0 % , suggesting that performing contrastive unlearning on even a small percentage of nodes ( 𝑘 ) significantly enhances the algorithm’s effectiveness. However, using higher values of 𝑘 yields similar performance with the added downside of increasing computational time (Right). E.3Comparing our sampling method to MEGU’s
Below, we compare the performance of Cognac while using our sampling method (described in Section 3.1.1) against its performance when using MEGU’s sampling method.
Table 7:Evaluating Cognac with a different strategy to identify affected nodes on Cora. We find that our strategy outperforms the MEGU’s, while also being 8 x faster. Both variants are hyperparameter-tuned for 100 runs. Method Acc rem
Acc aff Sampling Time (3160 nodes) Original 61.4 ± 0.00
40.0 ± 0.00 - Oracle 58.4 ± 0.00
69.2 ± 0.00 - Cognac (MEGU’s) 54.8 ± 0.00
48.7 ± 0.00
0.432 𝑠
Cognac (Ours) 56.6 ± 0.00
75.5 ± 0.04
0.054 𝐬 E.4Why does Acc aff sometimes reduce as identified manipulated entities increase?
We find an interesting trend that sometimes, as more of the manipulation set ( 𝑆 𝑚 ) is known and used as the deletion set ( 𝑆 𝑓 ) (going left to right in Figure 4), Acc aff reduces. This can seem counterintuitive, as one would expect the accuracy of affected classes to improve as more samples are used for unlearning. We hypothesize that unlearning a larger fraction of the manipulation set reduces Acc aff due to two factors that adversely affect the neighborhoods of the nodes removed, which typically have other nodes of the affected classes due to homophily. First, in the case of label manipulation, when we model it as node unlearning for consistency with prior work, we lose correct information about the graph structure. Second, when modifying the graph structure, i.e., removing some edges or nodes, changes the feature distribution of their neighboring nodes after the message passes, making it out of distribution for the learned GNN layers. The same rationale is why the test nodes are kept in the graph structure (without optimizing the task loss for them) during training (Kipf & Welling, 2017). We investigate this by adding an ablation where, in the unlearning of the label manipulation, instead of unlearning the whole node, we keep the structure, i.e., the node and connected edges, but unlearn the features and labels.
Table 8:Ablating node unlearning performance on label manipulation with and without unlinking. We report the accuracy on the affected classes Acc aff for unlearning the label manipulation on Cora, both when the full and a subset (25%) of the manipulated set is used for deletion. We find that not removing the structural information leads to a significant improvement in Acc aff , especially when more entities are deleted. This illustrates that the unlearning methods can achieve improved performance when precise information about the manipulated data is available. Method 0.25
1.00
Linked Unlinked Linked Unlinked
Oracle 73.0 ± 0.0
73.0 ± 0.0
73.0 ± 0.0
73.0 ± 0.0
Original 42.0 ± 0.0
42.0 ± 0.0
42.0 ± 0.0
42.0 ± 0.0
Cognac 64.8 ± 0.9
67.8 ± 3.2
77.2 ± 1.0
69.3 ± 1.3
GNNDelete 35.2 ± 2.5
50.2 ± 1.9
21.9 ± 4.5
30.2 ± 5.3
MEGU 40.8 ± 1.6
33.4 ± 0.4
41.2 ± 1.5
32.1 ± 1.3
SCRUB 45.7 ± 0.0
60.7 ± 0.0
41.1 ± 0.0
29.0 ± 0.0
AC ↯ DC 61.7 ± 0.0
58.8 ± 0.0
63.7 ± 0.0
54.3 ± 0.0
As observed in Table 8, retaining the node structure leads to large improvements in Acc aff when the deletion set is larger (the full set of manipulated entities), while not benefiting much when the deletion set is smaller. In the full manipulation set deletion setting, Cognac even slightly outperforms Oracle. This highlights how, unlike conventional node unlearning in graphs, removing the nodes is not always the best way to unlearn manipulations. They can simply be moved from the train set to the test set to still partake in message passing, so the task loss is not optimized over wrong labels.
Appendix FDetailed report of the experimental setup F.1Hyperparameter Tuning
We perform extensive hyperparameter tuning for all unlearning methods using Optuna (Akiba et al., 2019) with a TPESampler (Tree-structured Parzen Estimator) Algorithm. We ensure the hyperparameter ranges searched include any values specified by the methods. The optimization target is an average of Acc aff and Acc rem , computed on the validation set. We report averaged results across five seeds. Method-specific hyperparameter ranges and scatter plots across hyperparameters for each method are provided in Appendix F.1.
We perform hyperparameter tuning for each combination of attack, dataset, unlearning method, and the identified fraction of deletion set ( 𝑆 𝑓 ). The optimization target is an average of Acc aff and Acc rem , computed on the validation set. For each setting, we run 100 trials with hyperparameters selected using the TPESampler (Tree-structured Parzen Estimator) algorithm. In Figures 10, 11 and 12, we report Acc aff and Acc rem scores for each hyperparameter tuning trial. Across hyperparameter runs, existing graph-based unlearning methods, barring MEGU, vary drastically across different sets of hyperparameters. On the other hand, our proposed method Cognac and its ablations show consistently high scores across hyperparameters, showcasing Cognac’s robustness to hyperparameter tuning.
Figure 10:Hyperparameter runs for 𝑆 𝑓
1.0 on Amazon. Scores of various hyperparameter trial runs. The best hyperparameters are selected according to the run achieving the best value for the average of Acc rem and Acc aff . Figure 11:Hyperparameter runs for 𝑆 𝑓
1.0 on CS. Scores of various hyperparameter trial runs. The best hyperparameters are selected according to the run achieving the best value for the average of Acc rem and Acc aff . Figure 12:Hyperparameter runs for 𝑆 𝑓
1.0 on Cora. Scores of various hyperparameter trial runs. The best hyperparameters are selected according to the run achieving the best value for the average of Acc rem and Acc aff . F.2Unlearning Times
To ensure the practicality of inexact unlearning methods, they must achieve greater efficiency than retraining from scratch. As illustrated in Figure 13, Cognac demonstrates competitive efficiency compared to alternative methods, offering substantial speedups over retraining from scratch.
Measuring Unlearning Time. To simplify comparisons to just two axes, Acc aff , and Acc rem , we fix a maximum cutoff of time an unlearning method can take, as motivated by Maini et al. (2024). We chose this cutoff as 25% of the original model training time. We pick the best model checkpoint during training for each method, which could be achieved earlier than this. Average run times for each method reported in Figure 13 under Appendix F.2 show that Cognac’s efficiency is comparable to or better than baselines. All experiments were run on a machine with Intel Xeon CPUs and two dedicated RTX 5000 GPUs.
Figure 13:Time taken to unlearn. We report the time taken by each unlearning method for all datasets, in the setting where all manipulated samples are known for deletion. Cognac provides significant speedups over Retrain and has performance similar to other unlearning methods. Report Issue Report Issue for Selection Generated by L A T E xml Instructions for reporting errors
We are continuing to improve HTML versions of papers, and your feedback helps enhance accessibility and mobile support. To report errors in the HTML that will help us improve conversion and rendering, choose any of the methods listed below:
Click the "Report Issue" button. Open a report feedback form via keyboard, use "Ctrl + ?". Make a text selection and click the "Report Issue for Selection" button near your cursor. You can use Alt+Y to toggle on and Alt+Shift+Y to toggle off accessible reporting links at each section.
Our team has already identified the following issues. We appreciate your time reviewing and reporting rendering errors we may not have found yet. Your efforts will help us improve the HTML versions for all readers, because disability should not be a barrier to accessing research. Thank you for your continued support in championing open access for all.
Have a free development cycle? Help support accessibility at arXiv! Our collaborators at LaTeXML maintain a list of packages that need conversion, and welcome developer contributions.
Xet Storage Details
- Size:
- 95.9 kB
- Xet hash:
- 3f4a96165563cc602d2f4466df5ffd47cc978287f9a8d5de79c10c2021971903
Xet efficiently stores files, intelligently splitting them into unique chunks and accelerating uploads and downloads. More info.