Incorporating Dependencies in Spectral Kernels for Gaussian Processes
Gaussian processes (GPs) are an elegant Bayesian approach to model an unknown function. The choice of the kernel characterizes one’s assumption on how the unknown function autocovaries. It is a core aspect of a GP design, since the posterior distribution can significantly vary for different kernels....
Saved in:
Published in | Machine Learning and Knowledge Discovery in Databases pp. 565 - 581 |
---|---|
Main Authors | , , , |
Format | Book Chapter |
Language | English |
Published |
Cham
Springer International Publishing
2020
|
Series | Lecture Notes in Computer Science |
Subjects | |
Online Access | Get full text |
ISBN | 3030461467 9783030461461 |
ISSN | 0302-9743 1611-3349 |
DOI | 10.1007/978-3-030-46147-8_34 |
Cover
Loading…
Summary: | Gaussian processes (GPs) are an elegant Bayesian approach to model an unknown function. The choice of the kernel characterizes one’s assumption on how the unknown function autocovaries. It is a core aspect of a GP design, since the posterior distribution can significantly vary for different kernels. The spectral mixture (SM) kernel is derived by modelling a spectral density - the Fourier transform of a kernel - with a linear mixture of Gaussian components. As such, the SM kernel cannot model dependencies between components. In this paper we use cross convolution to model dependencies between components and derive a new kernel called Generalized Convolution Spectral Mixture (GCSM). Experimental analysis of GCSM on synthetic and real-life datasets indicates the benefit of modeling dependencies between components for reducing uncertainty and for improving performance in extrapolation tasks. |
---|---|
Bibliography: | Electronic supplementary materialThe online version of this chapter (10.1007/978-3-030-46147-8_34) contains supplementary material, which is available to authorized users. |
ISBN: | 3030461467 9783030461461 |
ISSN: | 0302-9743 1611-3349 |
DOI: | 10.1007/978-3-030-46147-8_34 |