Abstract
We introduce a new category of generative autoencoders called automodulators. These networks can faithfully reproduce individual real-world input images like regular autoencoders, but also generate a fused sample from an arbitrary combination of several such images, allowing instantaneous "style-mixing" and other new applications. An automodulator decouples the data flow of decoder operations from statistical properties thereof and uses the latent vector to modulate the former by the latter, with a principled approach for mutual disentanglement of decoder layers. Prior work has explored similar decoder architecture with GANs, but their focus has been on random sampling. A corresponding autoencoder could operate on real input images. For the first time, we show how to train such a general-purpose model with sharp outputs in high resolution, using novel training techniques, demonstrated on four image data sets. Besides style-mixing, we show state-of-the-art results in autoencoder comparison, and visual image quality nearly indistinguishable from state-of-the-art GANs. We expect the automodulator variants to become a useful building block for image applications and other data domains.
Original language | English |
---|---|
Title of host publication | Thirty-fourth Conference on Neural Information Processing Systems |
Publisher | Morgan Kaufmann Publishers |
Volume | 33 |
Publication status | Published - 2020 |
MoE publication type | A4 Article in a conference publication |
Event | Conference on Neural Information Processing Systems - Virtual, Vancouver, Canada Duration: 6 Dec 2020 → 12 Dec 2020 Conference number: 34 |
Publication series
Name | Advances in Neural Information Processing Systems |
---|---|
Publisher | Morgan Kaufmann Publishers |
Volume | 33 |
ISSN (Electronic) | 1049-5258 |
Conference
Conference | Conference on Neural Information Processing Systems |
---|---|
Abbreviated title | NeurIPS |
Country/Territory | Canada |
City | Vancouver |
Period | 06/12/2020 → 12/12/2020 |