TR2016-156
Coupled Generative Adversarial Nets
-
- "Coupled Generative Adversarial Nets", Advances in Neural Information Processing Systems (NIPS), December 2016.BibTeX TR2016-156 PDF
- @inproceedings{Liu2016dec,
- author = {Liu, Ming-Yu and Tuzel, C. Oncel and Sullivan, Alan},
- title = {Coupled Generative Adversarial Nets},
- booktitle = {Advances in Neural Information Processing Systems (NIPS)},
- year = 2016,
- month = dec,
- url = {https://www.merl.com/publications/TR2016-156}
- }
,
- "Coupled Generative Adversarial Nets", Advances in Neural Information Processing Systems (NIPS), December 2016.
-
Research Areas:
Abstract:
We propose the coupled generative adversarial nets (CoGAN) framework for generating pairs of corresponding images in two different domains. The framework consists of a pair of generative adversarial nets, each responsible for generating images in one domain. We show that by enforcing a simple weight-sharing constraint, the CoGAN learns to generate pairs of corresponding images without existence of any pairs of corresponding images in the two domains in the training set. In other words, the CoGAN learns a joint distribution of images in the two domains from images drawn separately from the marginal distributions of the individual domains. This is in contrast to the existing multi-modal generative models, which require corresponding images for training. We apply the CoGAN to several pair image generation tasks. For each task, the GoGAN learns to generate convincing pairs of corresponding images. We further demonstrate the applications of the CoGAN framework for the domain adaptation and cross-domain image generation tasks.