A principled approach to model validation in domain generalization
Domain generalization aims to learn a model with good generalization ability, that is, the learned model should not only perform well on several seen domains but also on unseen domains with different data distributions. State-of-the-art domain generalization methods typically train a representation...
Saved in:
Main Authors | , , , , |
---|---|
Format | Journal Article |
Language | English |
Published |
02.04.2023
|
Subjects | |
Online Access | Get full text |
Cover
Loading…
Summary: | Domain generalization aims to learn a model with good generalization ability,
that is, the learned model should not only perform well on several seen domains
but also on unseen domains with different data distributions. State-of-the-art
domain generalization methods typically train a representation function
followed by a classifier jointly to minimize both the classification risk and
the domain discrepancy. However, when it comes to model selection, most of
these methods rely on traditional validation routines that select models solely
based on the lowest classification risk on the validation set. In this paper,
we theoretically demonstrate a trade-off between minimizing classification risk
and mitigating domain discrepancy, i.e., it is impossible to achieve the
minimum of these two objectives simultaneously. Motivated by this theoretical
result, we propose a novel model selection method suggesting that the
validation process should account for both the classification risk and the
domain discrepancy. We validate the effectiveness of the proposed method by
numerical results on several domain generalization datasets. |
---|---|
DOI: | 10.48550/arxiv.2304.00629 |