Distill2Vec: Dynamic Graph Representation Learning with Knowledge Distillation

November 11, 2020 ยท Entered Twilight ยท ๐Ÿ› International Conference on Advances in Social Networks Analysis and Mining

๐ŸŒ… TWILIGHT: Old Age
Predates the code-sharing era โ€” a pioneer of its time

"No code URL or promise found in abstract"
"Derived repo from GitHub Pages (backfill)"

Evidence collected by the PWNC Scanner

Repo contents: .gitignore, .idea, LICENSE, README.md, _config.yml, data, docs, flags.py, models, raw_data, requirements.txt, run.sh, run_script.py, supplementary, train.py, utils

Authors Stefanos Antaris, Dimitrios Rafailidis arXiv ID 2011.05664 Category cs.LG: Machine Learning Cross-listed cs.AI Citations 3 Venue International Conference on Advances in Social Networks Analysis and Mining Repository https://github.com/stefanosantaris/Distill2Vec. โญ 3 Last Checked 1 month ago
Abstract
Dynamic graph representation learning strategies are based on different neural architectures to capture the graph evolution over time. However, the underlying neural architectures require a large amount of parameters to train and suffer from high online inference latency, that is several model parameters have to be updated when new data arrive online. In this study we propose Distill2Vec, a knowledge distillation strategy to train a compact model with a low number of trainable parameters, so as to reduce the latency of online inference and maintain the model accuracy high. We design a distillation loss function based on Kullback-Leibler divergence to transfer the acquired knowledge from a teacher model trained on offline data, to a small-size student model for online data. Our experiments with publicly available datasets show the superiority of our proposed model over several state-of-the-art approaches with relative gains up to 5% in the link prediction task. In addition, we demonstrate the effectiveness of our knowledge distillation strategy, in terms of number of required parameters, where Distill2Vec achieves a compression ratio up to 7:100 when compared with baseline approaches. For reproduction purposes, our implementation is publicly available at https://stefanosantaris.github.io/Distill2Vec.
Community shame:
Not yet rated
Community Contributions

Found the code? Know the venue? Think something is wrong? Let us know!

๐Ÿ“œ Similar Papers

In the same crypt โ€” Machine Learning