You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
This is the implementation of Latent Adversarial Training of Graph Convolutional Networks LATGCN by Hongwei Jin and Xinhua Zhang.
Motivation
maximize the perturbation on the latent layer
perturbation applied to all the nodes, due to the transductive learning property
improve the accuracy
reduce the success rate from adversarial attacks
Usage
Instal the required packages
pip install -r requirements.txt
Run the demo of nettack with vanilla GCN
python demo.py
run the demo of nettack with LATGCN
python demo.py
demo.py has a set of parameter to specify
--dataset: choose from
'citeseer', 'cora', 'cora_ml', 'pubmed', 'polblogs', 'dblp'
--train_share: specify the size for training
--reg: a flag to toggle the LATGCN
--eta: the norm constraint of noise applied to each node embedding
--gamma: regularizer factor
Note that without specify the reg flag, it is simply the vanilla GCN.
Primary Result
Performance with perturbation on cora dataset:
Success rate with perturbation on cora dataset:
Reference
The Nettack model is implemented by Daniel Zügner
The original GCN model is implemented by Thomas N Kipf
Contact
Feel free to create issue in the repo, I will try to answer as soon as I can.
Citation
@article{jin2019latent,
title={Latent Adversarial Training of Graph Convolutional Networks},
author={Jin, Hongwei and Zhang, Xinhua},
journal={ICML workshop - Learning and Reasoning with Graph-Structured Representations},
year={2019}
}