Data Distillation for Neural Network Potentials toward Foundational Dataset
November 09, 2023 ยท Declared Dead ยท ๐ arXiv.org
"No code URL or promise found in abstract"
Evidence collected by the PWNC Scanner
Authors
Gang Seob Jung, Sangkeun Lee, Jong Youl Choi
arXiv ID
2311.05407
Category
physics.comp-ph
Cross-listed
cs.LG,
physics.chem-ph
Citations
2
Venue
arXiv.org
Last Checked
1 month ago
Abstract
Machine learning (ML) techniques and atomistic modeling have rapidly transformed materials design and discovery. Specifically, generative models can swiftly propose promising materials for targeted applications. However, the predicted properties of materials through the generative models often do not match with calculated properties through ab initio calculations. This discrepancy can arise because the generated coordinates are not fully relaxed, whereas the many properties are derived from relaxed structures. Neural network-based potentials (NNPs) can expedite the process by providing relaxed structures from the initially generated ones. Nevertheless, acquiring data to train NNPs for this purpose can be extremely challenging as it needs to encompass previously unknown structures. This study utilized extended ensemble molecular dynamics (MD) to secure a broad range of liquid- and solid-phase configurations in one of the metallic systems, nickel. Then, we could significantly reduce them through active learning without losing much accuracy. We found that the NNP trained from the distilled data could predict different energy-minimized closed-pack crystal structures even though those structures were not explicitly part of the initial data. Furthermore, the data can be translated to other metallic systems (aluminum and niobium), without repeating the sampling and distillation processes. Our approach to data acquisition and distillation has demonstrated the potential to expedite NNP development and enhance materials design and discovery by integrating generative models.
Community Contributions
Found the code? Know the venue? Think something is wrong? Let us know!
๐ Similar Papers
In the same crypt โ physics.comp-ph
R.I.P.
๐ป
Ghosted
R.I.P.
๐ป
Ghosted
Deep Potential Molecular Dynamics: a scalable model with the accuracy of quantum mechanics
R.I.P.
๐ป
Ghosted
Heterogeneous Parallelization and Acceleration of Molecular Dynamics Simulations in GROMACS
R.I.P.
๐ป
Ghosted
By-passing the Kohn-Sham equations with machine learning
R.I.P.
๐ป
Ghosted
Machine Learning of coarse-grained Molecular Dynamics Force Fields
R.I.P.
๐ป
Ghosted
Towards Physics-informed Deep Learning for Turbulent Flow Prediction
Died the same way โ ๐ป Ghosted
R.I.P.
๐ป
Ghosted
Language Models are Few-Shot Learners
R.I.P.
๐ป
Ghosted
PyTorch: An Imperative Style, High-Performance Deep Learning Library
R.I.P.
๐ป
Ghosted
XGBoost: A Scalable Tree Boosting System
R.I.P.
๐ป
Ghosted