Contrastive Class-aware Adaptation for Domain Generalization
Domain generalization (DG) tackles the problem of learning a model that generalizes to data drawn from a target domain that was unseen during training. A major trend in this area consists of learning a domain-invariant representation by minimizing the discrepancy across multiple source domains. This strategy, however, does not apply to the challenging yet realistic single-source scenario. In this paper, in contrast to existing methods that focus on domain discrepancy, we exploit the fact that discrepancies also arise across samples from the same class. We therefore develop a unified framework for both multi-source and single-source DG that exploits contrastive learning to maximize the gap between samples from the same class, either from different domains or from the same one, while separating the samples from different classes. Our results on standard multi-source and single-source DG benchmark datasets demonstrate the benefits of our method over the state-of-the-art ones in both settings.
WOS:000897707604123
2022-01-01
New York
978-1-6654-9062-7
International Conference on Pattern Recognition
4871
4876
REVIEWED
EPFL
Event name | Event place | Event date |
Montreal, CANADA | Aug 21-25, 2022 | |