APS Logo

Evaluating the limits of the physics learned by a machine learning model

ORAL

Abstract

In machine learning for materials science, a model should learn relevant physics rather than arbitrary data representations. This has implications for the robustness, generalizability, and overall trustworthiness of model predictions. In this talk, the loss landscape of the model is used to probe the limits of the learned representation: we compare the geometry of the loss landscape between in-distribution and out-of-distribution prediction tasks. Discontinuities in the loss landscape may indicate where a model fails to generalize, and the perturbation scale of the discontinuities can quantify the robustness of the feature representations in the model weights. This method is demonstrated using graph neural networks trained for several out-of-distribution (OOD) generalization tasks from DFT materials datasets. These models, trained by omitting one or more elements, did not generalize to OOD samples containing the omitted chemistry. The reason why the models do not generalize is unclear, but the model robustness and trustworthiness can still be explored using the loss landscape of the model.

Publication: Evaluating the Limits of the Physics Learned by a Machine Learning Model by Dale, Li, DeCost, Hattrick-Simpers<br>Loss Landscape Analysis of Model Accuracy by Dale, Li, DeCost, Hattrick-Simpers<br>Trusted AI Toolkit for Scientists (TRAITS) by Dale, Yao, Hattrick-Simpers

Presenters

  • Ashley Dale

    University of Toronto

Authors

  • Ashley Dale

    University of Toronto

  • Kangming Li

    Acceleration Consortium, University of Toronto

  • Brian DeCost

    National Institute of Standards and Technology

  • Jason Hattrick-Simpers

    University of Toronto