Evaluating and Diagnosing Convergence for Stochastic Gradient Langevin Dynamics

Hernández, Sergio; López, Juan Luis

Abstract

Bayesian deep learning is a method to quantity uncertainty for black-box models such as deep learning architectures. Despite the huge success of these models for several tasks such as computer vision and natural language processing, the existing performance metrics rely on point estimates and cannot properly explain the uncertainty around the predictions. Knowing the model confidence is absolutely necessary to avoid over-optimistic decisions. Therefore, in this paper we propose a novel F score that takes into account the Leave-One-Out (LOO) performance using posterior samples from stochastic gradient Markov chain Monte Carlo algorithms. The new Fιoo metric is evaluated for two different Bayesian models under flat and hierarchical priors. The results for the MNIST data set indicate that the new approach avoids errors from finite sample estimates.

Más información

Fecha de publicación: 2021
Año de Inicio/Término: 15-19 Nov. 2021
Página de inicio: 1
Página final: 6
URL: https://doi:10.1109/SCCC54552.2021.9650355