APS Logo

Towards General and Robust Deep Learning at Scale

ORAL · Invited

Abstract

Modern AI systems have achieved impressive results in many specific domains, from image and speech recognition to natural language processing and mastering complex games such as chess and Go. However, they often remain inflexible, fragile and narrow, unable to continually adapt to a wide range of changing environments and novel tasks without "catastrophically forgetting" what they have learned before, to infer higher-order abstractions allowing for systematic generalization to out-of-distribution data, and to achieve the level of robustness necessary to "survive" various perturbations in their environment - a natural property of most biological intelligent systems, and a necessary property for successfully deploying AI systems in real-life applications. In this talk, I will provide a brief overview of our recent efforts towards making AI more broad (i.e., general/versatile) and more robust, focusing on continual learning, invariance and adversarial robustness. I will also emphasize the importance of developing an empirical science of AI behaviors, and focus on rapidly expanding field of neural scaling laws, which allow us to better compare and extrapolate behavior of various algorithms and models with increasing amounts of data, model size and computational resources.

Publication: Adversarial Feature Desensitization. Pouya Bashivan, Mojtaba Faramarzi, Touraj Laleh, Blake Aaron Richards and Irina Rish. In Proc of NeurIPS 2021.<br><br>Scaling Laws for the Few-Shot Adaptation of Pre-trained Image Classifiers. Gabriele Prato, Simon Guiroy, Ethan Caballero, Irina Rish and Sarath Chandar. arXiv preprint arXiv:2110.06990, 2021.<br><br>Sequoia: A Software Framework to Unify Continual Learning Research. Fabrice Normandin, Florian Golemo, Oleksiy Ostapenko, Pau Rodríguez, Matthew D. Riemer, Julio Hurtado, Khimya Khetarpal, Dominic Zhao, Ryan Lindeborg, Timothée Lesort, Laurent Charlin, Irina Rish and Massimo Caccia. arXiv preprint arXiv:2108.01005<br><br>Invariance Principle Meets Information Bottleneck for Out-of-Distribution Generalization<br>Kartik Ahuja, Ethan Caballero, Dinghuai Zhang, Jean-Christophe Gagnon-Audet, Yoshua Bengio, Ioannis Mitliagkas and Irina Rish. In Proc of NeurIPS 2021.<br><br>Online Fast Adaptation and Knowledge Accumulation: a New Approach to Continual Learning. Massimo Caccia, Pau Rodríguez, Oleksiy Ostapenko, Fabrice Normandin, Min Lin, Lucas Caccia, Issam H. Laradji, Irina Rish, Alexandre Lacoste, David Vázquez, Laurent Charlin. In Proc of NeurIPS-2020.

Presenters

  • Irina Rish

    University of Montreal, Mila - Quebec AI Institute

Authors

  • Irina Rish

    University of Montreal, Mila - Quebec AI Institute