Add surrogate extraction v0.2 for GNN models with class-based implementation #3
Add this suggestion to a batch that can be applied as a single commit.
This suggestion is invalid because no changes were made to the code.
Suggestions cannot be applied while the pull request is closed.
Suggestions cannot be applied while viewing a subset of changes.
Only one suggestion per line can be applied in a batch.
Add this suggestion to a batch that can be applied as a single commit.
Applying suggestions on deleted lines is not supported.
You must change the existing code in this line in order to create a valid suggestion.
Outdated suggestions cannot be applied.
This suggestion has been applied or marked resolved.
Suggestions cannot be applied from pending reviews.
Suggestions cannot be applied on multi-line comments.
Suggestions cannot be applied while the pull request is queued to merge.
Suggestion cannot be applied right now. Please check back later.
Surrogate Extraction Fidelity Results for GNN Models
This document presents the fidelity results of surrogate models when varying the number of attack nodes for three popular datasets: Cora, Citeseer, and PubMed.
Results Summary
Dataset: Cora
Dataset: Citeseer
Dataset: PubMed
Analysis
These results suggest that the PubMed dataset is more robust to surrogate extraction attacks compared to Cora and Citeseer.
Conclusion
The fidelity of surrogate models depends on the dataset characteristics and the number of attack nodes. The results can help in understanding the robustness of Graph Neural Networks (GNNs) to surrogate extraction attacks.