Statistical Mechanics of Kernel Regression and Wide Neural Networks
ORAL
Abstract
A theoretical understanding of generalization remains an open problem for many machine learning models, including deep networks where overparameterization leads to better performance. Here, we study this problem for kernel regression, which, besides being a popular machine learning method, also describes infinitely overparameterized neural networks. We develop an analytical theory of generalization in kernel regression using replica theory of statistical mechanics. This theory is applicable to any kernel and data distribution. Experiments with practical kernels including those arising from wide neural networks show perfect agreement with our theory. Further, our theory accurately predicts generalization performance of neural networks with modest widths. We provide an in-depth analysis of our analytical expression for kernel generalization. We show that kernel machines employ an inductive bias towards simple functions, preventing them to overfit the data. We characterize whether a kernel is compatible with a learning task in terms of sample efficiency. We identify a first order phase transition in our theory where more data may impair generalization when the task is noisy or not expressible by the kernel. Finally, we extend these results to out-of-distribution generalization.
–
Publication: Blake Bordelon, Abdulkadir Canatar, Cengiz Pehlevan, Spectrum Dependent Learning Curves in Kernel Regression and Wide Neural Networks, ICML, 2020<br><br>Abdulkadir Canatar, Blake Bordelon, Cengiz Pehlevan, Spectral bias and task-model alignment explain generalization in kernel regression and infinitely wide neural networks, Nature Communications, 2021<br><br>Abdulkadir Canatar, Blake Bordelon, Cengiz Pehlevan, Out-of-Distribution Generalization in Kernel Regression, NeurIPS, 2021
Presenters
-
Abdulkadir Canatar
Harvard University
Authors
-
Abdulkadir Canatar
Harvard University
-
Blake Bordelon
Harvard University
-
Cengiz Pehlevan
Harvard University