Prediction rigidities for data-driven chemistry
Creators
- 1. Laboratory of Computational Science and Modeling (COSMO), IMX, École Polytechnique Fédérale de Lausanne (EPFL), 1015 Lausanne, Switzerland
* Contact person
Description
The widespread application of machine learning (ML) to the chemical sciences is making it very important to understand how the ML models learn to correlate chemical structures with their properties, and what can be done to improve the training efficiency whilst guaranteeing interpretability and transferability. In this work, we demonstrate the wide utility of prediction rigidities, a family of metrics derived from the loss function, in understanding the robustness of ML model predictions. We show that the prediction rigidities allow the assessment of the model not only at the global level, but also on the local or the component-wise level at which the intermediate (e.g. atomic, body-ordered, or range-separated) predictions are made. We leverage these metrics to understand the learning behavior of different ML models, and to guide efficient dataset construction for model training. We finally implement the formalism for a ML model targeting a coarse-grained system to demonstrate the applicability of the prediction rigidities to an even broader class of atomistic modeling problems. This record contains all the data used for the analyses conducted in the associated work published in Faraday Discussions.
Files
File preview
files_description.md
All files
References
Journal reference (Paper in which the data and models were used for analysis) S. Chong, F. Bigi, F. Grasselli, P. Loche, M. Kellner, M. Ceriotti, Faraday Discussions (2024), doi: 10.1039/D4FD00101J