liu.seSearch for publications in DiVA
Operational message
There are currently operational disruptions. Troubleshooting is in progress.
Change search
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • oxford
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf
Semi-Supervised Learning for Few-Shot Image-to-Image Translation
Computer Vision Center, Universitat Autonoma de Barcelona, Spain.
Inception Institute of Artificial Intelligence, UAE.
Computer Vision Center, Universitat Autonoma de Barcelona, Spain.
Computer Vision Center, Universitat Autonoma de Barcelona, Spain.
Show others and affiliations
2020 (English)In: 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), IEEE , 2020, p. 4452-4461Conference paper, Published paper (Refereed)
Abstract [en]

In the last few years, unpaired image-to-image translation has witnessed Remarkable progress. Although the latest methods are able to generate realistic images, they crucially rely on a large number of labeled images. Recently, some methods have tackled the challenging setting of few-shot image-to-image ranslation, reducing the labeled data requirements for the target domain during inference. In this work, we go one step further and reduce the amount of required labeled data also from the source domain during training. To do so, we propose applying semi-supervised learning via a noise-tolerant pseudo-labeling procedure. We also apply a cycle consistency constraint to further exploit the information from unlabeled images, either from the same dataset or external. Additionally, we propose several structural modifications to facilitate the image translation task under these circumstances. Our semi-supervised method for few-shot image translation, called SEMIT, achieves excellent results on four different datasets using as little as 10% of the source labels, and matches the performance of the main fully-supervised competitor using only 20% labeled data. Our code and models are made public at: https://github.com/yaxingwang/SEMIT.

Place, publisher, year, edition, pages
IEEE , 2020. p. 4452-4461
Series
Computer Society Conference on Computer Vision and Pattern Recognition, ISSN 2575-7075
Keywords [en]
Training;Task analysis;Labeling;Semisupervised learning;Head;Noise measurement;Entropy
National Category
Computer graphics and computer vision
Identifiers
URN: urn:nbn:se:liu:diva-168122DOI: 10.1109/CVPR42600.2020.00451ISI: 000620679504073ISBN: 978-1-7281-7168-5 (electronic)OAI: oai:DiVA.org:liu-168122DiVA, id: diva2:1458539
Conference
Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 13-19 June 2020
Note

Funding:  [TIN2016-79717-R]

Available from: 2020-08-17 Created: 2020-08-17 Last updated: 2025-02-07

Open Access in DiVA

No full text in DiVA

Other links

Publisher's full text

Authority records

Khan, Fahad Shahbaz

Search in DiVA

By author/editor
Khan, Fahad Shahbaz
By organisation
Computer VisionFaculty of Science & Engineering
Computer graphics and computer vision

Search outside of DiVA

GoogleGoogle Scholar

doi
isbn
urn-nbn

Altmetric score

doi
isbn
urn-nbn
Total: 104 hits
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • oxford
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf