TriGAN: image-to-image translation for multi-source domain adaptation

  • Subhankar Roy
  • , Aliaksandr Siarohin
  • , Enver Sangineto* (Corresponding Author)
  • , Nicu Sebe
  • , Elisa Ricci
  • *Corresponding author for this work

Research output: Contribution to journalArticlepeer-review

Abstract

Most domain adaptation methods consider the problem of transferring knowledge to the target domain from a single-source dataset. However, in practical applications, we typically have access to multiple sources. In this paper we propose the first approach for multi-source domain adaptation (MSDA) based on generative adversarial networks. Our method is inspired by the observation that the appearance of a given image depends on three factors: the domain, the style (characterized in terms of low-level features variations) and the content. For this reason, we propose to project the source image features onto a space where only the dependence from the content is kept, and then re-project this invariant representation onto the pixel space using the target domain and style. In this way, new labeled images can be generated which are used to train a final target classifier. We test our approach using common MSDA benchmarks, showing that it outperforms state-of-the-art methods.

Original languageEnglish
Article number41
Number of pages12
JournalMachine Vision and Applications
Volume32
Issue number1
Early online date19 Jan 2021
DOIs
Publication statusPublished - Jan 2021

Bibliographical note

Funding: Open Access funding provided by Universitá degli Studi di Trento

Data Availability Statement

Supplementary Information: The online version contains supplementary material available at https://doi.org/10.1007/s00138-020-01164-4.

Keywords

  • Generative adversarial network
  • Image classification
  • Image-to-image translation
  • Unsupervised domain adaptation

Fingerprint

Dive into the research topics of 'TriGAN: image-to-image translation for multi-source domain adaptation'. Together they form a unique fingerprint.

Cite this