Machine-learned metrics for predicting the likelihood of success in materials discovery
November 25, 2019 Β· Declared Dead Β· π npj Computational Materials
"No code URL or promise found in abstract"
Evidence collected by the PWNC Scanner
Authors
Yoolhee Kim, Edward Kim, Erin Antono, Bryce Meredig, Julia Ling
arXiv ID
1911.11201
Category
cond-mat.mtrl-sci
Cross-listed
cs.LG,
physics.comp-ph
Citations
27
Venue
npj Computational Materials
Last Checked
1 month ago
Abstract
Materials discovery is often compared to the challenge of finding a needle in a haystack. While much work has focused on accurately predicting the properties of candidate materials with machine learning (ML), which amounts to evaluating whether a given candidate is a piece of straw or a needle, less attention has been paid to a critical question: Are we searching in the right haystack? We refer to the haystack as the design space for a particular materials discovery problem (i.e. the set of possible candidate materials to synthesize), and thus frame this question as one of design space selection. In this paper, we introduce two metrics, the Predicted Fraction of Improved Candidates (PFIC), and the Cumulative Maximum Likelihood of Improvement (CMLI), which we demonstrate can identify discovery-rich and discovery-poor design spaces, respectively. Using CMLI and PFIC together to identify optimal design spaces can significantly accelerate ML-driven materials discovery.
Community Contributions
Found the code? Know the venue? Think something is wrong? Let us know!
π Similar Papers
In the same crypt β cond-mat.mtrl-sci
R.I.P.
π»
Ghosted
R.I.P.
π»
Ghosted
Microstructure Representation and Reconstruction of Heterogeneous Materials via Deep Belief Network for Computational Material Design
R.I.P.
π»
Ghosted
Deep learning and the SchrΓΆdinger equation
R.I.P.
π»
Ghosted
MatterGen: a generative model for inorganic materials design
R.I.P.
π»
Ghosted
Polymer Informatics with Multi-Task Learning
R.I.P.
π»
Ghosted
Inorganic Materials Synthesis Planning with Literature-Trained Neural Networks
Died the same way β π» Ghosted
R.I.P.
π»
Ghosted
Language Models are Few-Shot Learners
R.I.P.
π»
Ghosted
PyTorch: An Imperative Style, High-Performance Deep Learning Library
R.I.P.
π»
Ghosted
XGBoost: A Scalable Tree Boosting System
R.I.P.
π»
Ghosted