Attending to Graph Transformers

February 08, 2023 Β· Entered Twilight Β· πŸ› Trans. Mach. Learn. Res.

πŸ’€ TWILIGHT: Eternal Rest
Repo abandoned since publication

Repo contents: .gitattributes, .gitignore, LICENSE, README.md, configs, graphgps, main.py, run, setup.py, tests, unittests

Authors Luis Müller, Mikhail Galkin, Christopher Morris, Ladislav RampÑőek arXiv ID 2302.04181 Category cs.LG: Machine Learning Cross-listed cs.AI, cs.NE Citations 126 Venue Trans. Mach. Learn. Res. Repository https://github.com/luis-mueller/probing-graph-transformers ⭐ 92 Last Checked 1 month ago
Abstract
Recently, transformer architectures for graphs emerged as an alternative to established techniques for machine learning with graphs, such as (message-passing) graph neural networks. So far, they have shown promising empirical results, e.g., on molecular prediction datasets, often attributed to their ability to circumvent graph neural networks' shortcomings, such as over-smoothing and over-squashing. Here, we derive a taxonomy of graph transformer architectures, bringing some order to this emerging field. We overview their theoretical properties, survey structural and positional encodings, and discuss extensions for important graph classes, e.g., 3D molecular graphs. Empirically, we probe how well graph transformers can recover various graph properties, how well they can deal with heterophilic graphs, and to what extent they prevent over-squashing. Further, we outline open challenges and research direction to stimulate future work. Our code is available at https://github.com/luis-mueller/probing-graph-transformers.
Community shame:
Not yet rated
Community Contributions

Found the code? Know the venue? Think something is wrong? Let us know!

πŸ“œ Similar Papers

In the same crypt β€” Machine Learning