ComboGAN: Unrestrained Scalability for Image Domain Translation

December 19, 2017 ยท Entered Twilight ยท ๐Ÿ› 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW)

๐ŸŒ… TWILIGHT: Old Age
Predates the code-sharing era โ€” a pioneer of its time

"Last commit was 7.0 years ago (โ‰ฅ5 year threshold)"

Evidence collected by the PWNC Scanner

Repo contents: .gitignore, LICENSE, README.md, data, datasets, img, models, options, scripts, test.py, train.py, util

Authors Asha Anoosheh, Eirikur Agustsson, Radu Timofte, Luc Van Gool arXiv ID 1712.06909 Category cs.CV: Computer Vision Citations 209 Venue 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW) Repository https://github.com/AAnoosheh/ComboGAN โญ 138 Last Checked 2 months ago
Abstract
This year alone has seen unprecedented leaps in the area of learning-based image translation, namely CycleGAN, by Zhu et al. But experiments so far have been tailored to merely two domains at a time, and scaling them to more would require an quadratic number of models to be trained. And with two-domain models taking days to train on current hardware, the number of domains quickly becomes limited by the time and resources required to process them. In this paper, we propose a multi-component image translation model and training scheme which scales linearly - both in resource consumption and time required - with the number of domains. We demonstrate its capabilities on a dataset of paintings by 14 different artists and on images of the four different seasons in the Alps. Note that 14 data groups would need (14 choose 2) = 91 different CycleGAN models: a total of 182 generator/discriminator pairs; whereas our model requires only 14 generator/discriminator pairs.
Community shame:
Not yet rated
Community Contributions

Found the code? Know the venue? Think something is wrong? Let us know!

๐Ÿ“œ Similar Papers

In the same crypt โ€” Computer Vision