Loading Events

AI Seminar

A Defense of (Empirical) Neural Tangent Kernels

Danica SutherlandAssistant ProfessorUniversity of British Columbia
3725 Beyster BuildingMap


Neural tangent kernels, recently the hottest thing in deep learning theory, seem not to describe the full training process of practical neural networks very accurately. Even so, evidence is mounting that finite “empirical” neural tangent kernels can be highly effective for understanding real networks, especially “locally.” We discuss their use as a tool in understanding phenomena that help us guide the training process, and for practical purposes in approximate “look-ahead” criteria for active learning. We also present a new, theoretically-justified approximation to the empirical NTK which can save several orders of magnitude of computational cost, without substantially harming accuracy.


Danica Sutherland is an Assistant Professor in computer science at the University of British Columbia, and a Canada CIFAR AI Chair at Amii. She did her PhD at Carnegie Mellon University, a postdoc at University College London’s Gatsby unit, and was a research assistant professor at TTI-Chicago. Her research focuses on representation learning (particularly in forms that integrate kernel methods with deep learning), statistical learning theory, and understanding differences between probability distributions.


https://umich.zoom.us/j/92216884113 (password: UMichAI)


AI Lab

Faculty Host

Wei Hu