R.I.P.
๐ป
Ghosted
On the Initialization of Graph Neural Networks
December 05, 2023 ยท Entered Twilight ยท ๐ International Conference on Machine Learning
Repo contents: .gitignore, .vscode, README.md, comp1.py, comp10.py, comp11.py, comp2.py, comp3.py, comp4.py, comp5.py, comp6.py, comp7.py, comp8.py, comp9.py, list.md, tests, utils.py, utils_pyg.py
Authors
Jiahang Li, Yakun Song, Xiang Song, David Paul Wipf
arXiv ID
2312.02622
Category
cs.LG: Machine Learning
Cross-listed
cs.AI
Citations
10
Venue
International Conference on Machine Learning
Repository
https://github.com/LspongebobJH/virgo_icml2023
โญ 8
Last Checked
1 month ago
Abstract
Graph Neural Networks (GNNs) have displayed considerable promise in graph representation learning across various applications. The core learning process requires the initialization of model weight matrices within each GNN layer, which is typically accomplished via classic initialization methods such as Xavier initialization. However, these methods were originally motivated to stabilize the variance of hidden embeddings and gradients across layers of Feedforward Neural Networks (FNNs) and Convolutional Neural Networks (CNNs) to avoid vanishing gradients and maintain steady information flow. In contrast, within the GNN context classical initializations disregard the impact of the input graph structure and message passing on variance. In this paper, we analyze the variance of forward and backward propagation across GNN layers and show that the variance instability of GNN initializations comes from the combined effect of the activation function, hidden dimension, graph structure and message passing. To better account for these influence factors, we propose a new initialization method for Variance Instability Reduction within GNN Optimization (Virgo), which naturally tends to equate forward and backward variances across successive layers. We conduct comprehensive experiments on 15 datasets to show that Virgo can lead to superior model performance and more stable variance at initialization on node classification, link prediction and graph classification tasks. Codes are in https://github.com/LspongebobJH/virgo_icml2023.
Community Contributions
Found the code? Know the venue? Think something is wrong? Let us know!
๐ Similar Papers
In the same crypt โ Machine Learning
R.I.P.
๐ป
Ghosted
XGBoost: A Scalable Tree Boosting System
R.I.P.
๐ป
Ghosted
Batch Normalization: Accelerating Deep Network Training by Reducing Internal Covariate Shift
R.I.P.
๐ป
Ghosted
Semi-Supervised Classification with Graph Convolutional Networks
R.I.P.
๐ป
Ghosted
Proximal Policy Optimization Algorithms
R.I.P.
๐ป
Ghosted