Graph Adversarial Training: Dynamically Regularizing Based on Graph Structure

February 20, 2019 ยท Entered Twilight ยท ๐Ÿ› IEEE Transactions on Knowledge and Data Engineering

๐ŸŒ… TWILIGHT: Old Age
Predates the code-sharing era โ€” a pioneer of its time

"Last commit was 5.0 years ago (โ‰ฅ5 year threshold)"

Evidence collected by the PWNC Scanner

Repo contents: .gvat_citation.py.swp, README.md, __pycache__, data, gvat_citation.py, inits.py, layers.py, metrics.py, my_utils.py, vat_citation.py

Authors Fuli Feng, Xiangnan He, Jie Tang, Tat-Seng Chua arXiv ID 1902.08226 Category cs.LG: Machine Learning Cross-listed cs.SI, stat.ML Citations 252 Venue IEEE Transactions on Knowledge and Data Engineering Repository https://github.com/fulifeng/GraphAT โญ 20 Last Checked 1 month ago
Abstract
Recent efforts show that neural networks are vulnerable to small but intentional perturbations on input features in visual classification tasks. Due to the additional consideration of connections between examples (\eg articles with citation link tend to be in the same class), graph neural networks could be more sensitive to the perturbations, since the perturbations from connected examples exacerbate the impact on a target example. Adversarial Training (AT), a dynamic regularization technique, can resist the worst-case perturbations on input features and is a promising choice to improve model robustness and generalization. However, existing AT methods focus on standard classification, being less effective when training models on graph since it does not model the impact from connected examples. In this work, we explore adversarial training on graph, aiming to improve the robustness and generalization of models learned on graph. We propose Graph Adversarial Training (GraphAT), which takes the impact from connected examples into account when learning to construct and resist perturbations. We give a general formulation of GraphAT, which can be seen as a dynamic regularization scheme based on the graph structure. To demonstrate the utility of GraphAT, we employ it on a state-of-the-art graph neural network model --- Graph Convolutional Network (GCN). We conduct experiments on two citation graphs (Citeseer and Cora) and a knowledge graph (NELL), verifying the effectiveness of GraphAT which outperforms normal training on GCN by 4.51% in node classification accuracy. Codes are available via: https://github.com/fulifeng/GraphAT.
Community shame:
Not yet rated
Community Contributions

Found the code? Know the venue? Think something is wrong? Let us know!

๐Ÿ“œ Similar Papers

In the same crypt โ€” Machine Learning