Graph Transformer Networks for Accurate Band Structure Prediction: An End-to-End Approach
November 25, 2024 Β· Declared Dead Β· π arXiv.org
"No code URL or promise found in abstract"
Evidence collected by the PWNC Scanner
Authors
Weiyi Gong, Tao Sun, Hexin Bai, Jeng-Yuan Tsai, Haibin Ling, Qimin Yan
arXiv ID
2411.16483
Category
cond-mat.mtrl-sci
Cross-listed
cs.LG
Citations
4
Venue
arXiv.org
Last Checked
1 month ago
Abstract
Predicting electronic band structures from crystal structures is crucial for understanding structure-property correlations in materials science. First-principles approaches are accurate but computationally intensive. Recent years, machine learning (ML) has been extensively applied to this field, while existing ML models predominantly focus on band gap predictions or indirect band structure estimation via solving predicted Hamiltonians. An end-to-end model to predict band structure accurately and efficiently is still lacking. Here, we introduce a graph Transformer-based end-to-end approach that directly predicts band structures from crystal structures with high accuracy. Our method leverages the continuity of the k-path and treat continuous bands as a sequence. We demonstrate that our model not only provides accurate band structure predictions but also can derive other properties (such as band gap, band center, and band dispersion) with high accuracy. We verify the model performance on large and diverse datasets.
Community Contributions
Found the code? Know the venue? Think something is wrong? Let us know!
π Similar Papers
In the same crypt β cond-mat.mtrl-sci
R.I.P.
π»
Ghosted
R.I.P.
π»
Ghosted
Microstructure Representation and Reconstruction of Heterogeneous Materials via Deep Belief Network for Computational Material Design
R.I.P.
π»
Ghosted
Deep learning and the SchrΓΆdinger equation
R.I.P.
π»
Ghosted
MatterGen: a generative model for inorganic materials design
R.I.P.
π»
Ghosted
Polymer Informatics with Multi-Task Learning
R.I.P.
π»
Ghosted
Inorganic Materials Synthesis Planning with Literature-Trained Neural Networks
Died the same way β π» Ghosted
R.I.P.
π»
Ghosted
Language Models are Few-Shot Learners
R.I.P.
π»
Ghosted
PyTorch: An Imperative Style, High-Performance Deep Learning Library
R.I.P.
π»
Ghosted
XGBoost: A Scalable Tree Boosting System
R.I.P.
π»
Ghosted