Effective Theory of Deep Neural Networks
ORAL · Invited
Abstract
Large neural networks perform extremely well in practice, providing the backbone of modern machine learning. The goal of this talk is to provide a blueprint for theoretically analyzing these large models from first principles. In particular, we'll overview how the statistics and dynamics of deep neural networks drastically simplify at large width and become analytically tractable. In so doing, we'll see that the idealized infinite-width limit is too simple to capture several important aspects of deep learning such as representation learning. To address them, we'll step beyond the idealized limit and systematically incorporate finite-width corrections.
–
Publication: https://www.cambridge.org/core/books/principles-of-deep-learning-theory/3E566F65026D6896DC814A8C31EF3B4C
Presenters
-
Sho Yaida
Facebook AI Research
Authors
-
Sho Yaida
Facebook AI Research