Rethinking the Truly Unsupervised Image-to-Image Translation

Cited 15 time in webofscience Cited 0 time in scopus
  • Hit : 88
  • Download : 0
Every recent image-to-image translation model inherently requires either image-level (i.e. input-output pairs) or set-level (i.e. domain labels) supervision. However, even set-level supervision can be a severe bottleneck for data collection in practice. In this paper, we tackle image-to-image translation in a fully unsupervised setting, i.e., neither paired images nor domain labels. To this end, we propose a truly unsupervised image-to-image translation model (TUNIT) that simultaneously learns to separate image domains and translates input images into the estimated domains. Experimental results show that our model achieves comparable or even better performance than the set-level supervised model trained with full labels, generalizes well on various datasets, and is robust against the choice of hyperparameters (e.g. the preset number of pseudo domains). Furthermore, TUNIT can be easily extended to semi-supervised learning with a few labeled data.
Publisher
Institute of Electrical and Electronics Engineers Inc.
Issue Date
2021-10-12
Language
English
Citation

18th IEEE/CVF International Conference on Computer Vision, ICCV 2021, pp.14134 - 14143

DOI
10.1109/ICCV48922.2021.01389
URI
http://hdl.handle.net/10203/298044
Appears in Collection
AI-Conference Papers(학술대회논문)
Files in This Item
There are no files associated with this item.
This item is cited by other documents in WoS
⊙ Detail Information in WoSⓡ Click to see webofscience_button
⊙ Cited 15 items in WoS Click to see citing articles in records_button

qr_code

  • mendeley

    citeulike


rss_1.0 rss_2.0 atom_1.0