Deep learning uses the language and tools of statistics and classical machine learning, including empirical and population losses and optimizing a hypothesis on a training set. But it uses these tools in regimes where they should not be applicable: the optimization task is non-convex, models are often large enough to overfit, and the training and deployment tasks can radically differ. In this talk I will survey the relation between deep learning and statistics. In particular we will discuss recent works supporting the emerging intuition that deep learning is closer in some aspects to human learning than to classical statistics. Rather than estimating quantities from samples, deep neural nets develop broadly applicable representations and skills through their training. The talk will not assume background knowledge in artificial intelligence or deep learning.
TILOS Fireside Chat on "Theory in the age of modern AI", which will be a conversation led by TILOS team members: Misha Belkin (UCSD), Arya Mazumdar (moderator, UCSD), Tara Javidi (UCSD), Visheeth Vishnoi (Yale U). The focus will be on the implications to, and the roles played by theory, in modern AI (especially with the recent exciting […]