Meta-learning to calibrate Gaussian processes with deep kernels for regression uncertainty estimation

Although Gaussian processes (GPs) with deep kernels have been successfully used for meta-learning in regression tasks, its uncertainty estimation performance can be poor. We propose a meta-learning method for calibrating deep kernel GPs for improving regression uncertainty estimation performance wit...

Full description

Saved in:
Bibliographic Details
Published inNeurocomputing (Amsterdam) Vol. 579; p. 127441
Main Authors Iwata, Tomoharu, Kumagai, Atsutoshi
Format Journal Article
LanguageEnglish
Published Elsevier B.V 28.04.2024
Subjects
Online AccessGet full text

Cover

Loading…
More Information
Summary:Although Gaussian processes (GPs) with deep kernels have been successfully used for meta-learning in regression tasks, its uncertainty estimation performance can be poor. We propose a meta-learning method for calibrating deep kernel GPs for improving regression uncertainty estimation performance with a limited number of training data. The proposed method meta-learns how to calibrate uncertainty using data from various tasks by minimizing the test expected calibration error, and uses the knowledge for unseen tasks. We design our model such that the adaptation and calibration for each task can be performed without iterative procedures, which enables effective meta-learning. In particular, a task-specific uncalibrated output distribution is modeled by a GP with a task-shared encoder network, and it is transformed to a calibrated one using a cumulative distribution function of a task-specific Gaussian mixture model (GMM). By integrating the GP and GMM into our neural network-based model, we can meta-learn the model parameters in an end-to-end fashion. Our experiments demonstrate that the proposed method improves uncertainty estimation performance while keeping high regression performance compared with the existing methods using real-world datasets in few-shot settings.
ISSN:0925-2312
1872-8286
DOI:10.1016/j.neucom.2024.127441