GPS++: An Optimised Hybrid MPNN/Transformer for Molecular Property Prediction

November 18, 2022 Β· Entered Twilight Β· πŸ› arXiv.org

πŸ’€ TWILIGHT: Eternal Rest
Repo abandoned since publication

Repo contents: .gitignore, .style.yapf, GPSplusplus.pdf, LICENSE, Makefile, NOTICE, OGB_paper_diagram.png, README.md, argparser.py, configs, conftest.py, custom_callbacks.py, data_utils, inference.py, model, notebook_inference.ipynb, notebook_training.ipynb, notebook_utils.py, pcqm4mv2-cross_val_splits, pipeline, plotting.py, pytest.ini, requirements-dev.txt, requirements.txt, run_training.py, static_ops, tests, utils.py, xpu.py

Authors Dominic Masters, Josef Dean, Kerstin Klaser, Zhiyi Li, Sam Maddrell-Mander, Adam Sanders, Hatem Helal, Deniz Beker, Ladislav RampÑőek, Dominique Beaini arXiv ID 2212.02229 Category q-bio.QM Cross-listed cs.LG Citations 30 Venue arXiv.org Repository https://github.com/graphcore/ogb-lsc-pcqm4mv2 ⭐ 79 Last Checked 1 month ago
Abstract
This technical report presents GPS++, the first-place solution to the Open Graph Benchmark Large-Scale Challenge (OGB-LSC 2022) for the PCQM4Mv2 molecular property prediction task. Our approach implements several key principles from the prior literature. At its core our GPS++ method is a hybrid MPNN/Transformer model that incorporates 3D atom positions and an auxiliary denoising task. The effectiveness of GPS++ is demonstrated by achieving 0.0719 mean absolute error on the independent test-challenge PCQM4Mv2 split. Thanks to Graphcore IPU acceleration, GPS++ scales to deep architectures (16 layers), training at 3 minutes per epoch, and large ensemble (112 models), completing the final predictions in 1 hour 32 minutes, well under the 4 hour inference budget allocated. Our implementation is publicly available at: https://github.com/graphcore/ogb-lsc-pcqm4mv2.
Community shame:
Not yet rated
Community Contributions

Found the code? Know the venue? Think something is wrong? Let us know!

πŸ“œ Similar Papers

In the same crypt β€” q-bio.QM