ISF-GAN: An Implicit Style Function for High-Resolution Image-to-Image Translation

Yahui Liu, Yajing Chen, Linchao Bao, Nicu Sebe, Bruno Lepri, Marco De Nadai

Research output: Contribution to journalArticlepeer-review

Abstract

Recently, there has been an increasing interest in image editing methods that employ pre-trained unconditional image generators (e.g., StyleGAN). However, applying these methods to translate images to multiple visual domains remains challenging. Existing works do not often preserve the domain-invariant part of the image (e.g., the identity in human face translations), or they do not usually handle multiple domains or allow for multi-modal translations. This work proposes an implicit style function (ISF) to straightforwardly achieve multi-modal and multi-domain image-to-image translation from pre-trained unconditional generators. The ISF manipulates the semantics of a latent code to ensure that the image generated from the manipulated code lies in the desired visual domain. Our human faces and animal image manipulations show significantly improved results over the baselines. Our model enables cost-effective multi-modal unsupervised image-to-image translations at high resolution using pre-trained unconditional GANs. The code and data are available at: https://github.com/yhlleo/stylegan-mmuit.
Original languageEnglish
Article number9735294
Pages (from-to)3343-3353
Number of pages11
JournalIEEE Transactions on Multimedia
Volume25
Early online date15 Mar 2022
DOIs
Publication statusPublished - 8 Aug 2023

Keywords

  • Codes
  • Semantics
  • Image resolution
  • Generators
  • Training
  • Task analysis
  • Visualization

Fingerprint

Dive into the research topics of 'ISF-GAN: An Implicit Style Function for High-Resolution Image-to-Image Translation'. Together they form a unique fingerprint.

Cite this