CS Machine Learning Seminar: Plex—Towards Reliability using Pretrained Large Model Extensions

Tuesday, September 20, 2022
3:30 p.m.
Online presentation
Soheil Feizi
sfeizi@umd.edu

CS Machine Learning Seminar

Plex: Towards Reliability using Pretrained Large Model Extensions

Balaji Lakshminarayanan
Google Brain

https://umd.zoom.us/j/95197245230?pwd=cDRlVWRVeXBHcURGQkptSHpIS0VGdz09

Password: 828w
 
Abstract
A recent trend in artificial intelligence is the use of pretrained models for language and vision tasks, which have achieved extraordinary performance but also puzzling failures. Probing these models' abilities in diverse ways is therefore critical to the field. I will talk about our recent work exploring the reliability of models, where we define a reliable model as one that not only achieves strong predictive performance but also performs well consistently over many decision-making tasks involving uncertainty (e.g., selective prediction, open set recognition, calibration under shift), robust generalization (e.g., accuracy and log-likelihood on in- and out-of-distribution datasets), and adaptation (e.g., active learning, few-shot uncertainty). Plex builds on our work on scalable building blocks for probabilistic deep learning such as Gaussian process last-layer and efficient variants of deep ensembles. We show that Plex improves the state-of-the-art across reliability tasks, and simplifies the traditional protocol as it improves the out-of-the-box performance and does not require designing scores or tuning the model for each task. 
 
References
Plex: Towards Reliability using Pretrained Large Model Extensions https://arxiv.org/abs/2207.07411  (blog: https://ai.googleblog.com/2022/07/towards-reliability-in-deep-learning.html)
Practical Uncertainty Estimation & Out-of-Distribution Robustness in Deep Learning (NeurIPS'2020 tutorial slides)
A Simple Approach to Improve Single-Model Deep Uncertainty via Distance-Awareness https://arxiv.org/abs/2205.00403
Deep Ensembles: A Loss Landscape Perspective https://arxiv.org/abs/1912.02757
 

Bio
Balaji Lakshminarayanan is a staff research scientist at Google Brain. His recent research is focused on probabilistic deep learning, specifically, uncertainty estimation, out-of-distribution robustness and applications. Before joining Google Brain, he was a research scientist at DeepMind. He received his PhD from the Gatsby Unit, University College London where he worked with Yee Whye Teh. He has co-organized several workshops on "Uncertainty and Robustness in deep learning" and served as Area Chair for NeurIPS, ICML, ICLR and AISTATS.

remind we with google calendar

 

June 2025

SU MO TU WE TH FR SA
1 2 3 4 5 6 7
8 9 10 11 12 13 14
15 16 17 18 19 20 21
22 23 24 25 26 27 28
29 30 1 2 3 4 5
Submit an Event