APS Logo

Soft Matter Physics for Machine Learning: Dynamical loss functions

ORAL

Abstract

The neural network architectures, loss functions or optimizing protocols that are used in deep learning often stem from laborious trial-and-error design. This has triggered great interest in improving theoretical understanding of the connection between the structure of the loss function landscape and the performance of the optimizing protocol, or algorithm. So far, most effort has focused on improving the algorithm (e.g. stochastic gradient descent). We take a different approach by exploring new loss functions. In particular, we explore the effect of dynamical loss functions that change during training. Preliminary results show that this new approach can outperform results obtained with static loss functions for particular cases. We use the Hessian and Neural Tangent Kernel spectrums to understand how topographical changes of the loss function landscape can improve learning.

Presenters

  • Miguel Ruiz Garcia

    Technical University of Madrid, University of Pennsylvania

Authors

  • Miguel Ruiz Garcia

    Technical University of Madrid, University of Pennsylvania

  • Ge Zhang

    University of Pennsylvania

  • Sam Schoenholz

    Google Brain

  • Andrea Liu

    University of Pennsylvania, Department of Physics and Astronomy, University of Pennsylvania