Paper [19] on Similar Typed Data

Paper [19] on Similar Typed Data

Diverse Image Generation via Self-Conditioned GANs Steven Liu1 Tongzhou Wang1 David Bau1 Jun-Yan Zhu2 Antonio Torralba1 1MIT CSAIL 2Adobe Research Real Generated Real Generated Real Generated Real Generated Places365 Cluster 03 Cluster 61 Cluster 75 Cluster 99 ImageNet Cluster 19 Cluster 59 Cluster 60 Cluster 66 Figure 1: Our proposed self-conditioned GAN model learns to perform clustering and image synthesis simultaneously. The model training requires no manual annotation of object classes. Here, we visualize several discovered clusters for both Places365 (top) and ImageNet (bottom). For each cluster, we show both real images and the generated samples conditioned on the cluster index. Abstract 1. Introduction We introduce a simple but effective unsupervised method Despite the remarkable progress of Generative Adversar- for generating realistic and diverse images. We train a class- ial Networks (GANs) [16, 7, 25], there remains a signifi- conditional GAN model without using manually annotated cant gap regarding the quality and diversity between class- class labels. Instead, our model is conditional on labels conditional GANs trained on labeled data, and unconditional automatically derived from clustering in the discriminator’s GANs trained without any labels in a fully unsupervised set- feature space. Our clustering step automatically discov- ting [38, 34]. This problem reflects the challenge of mode ers diverse modes, and explicitly requires the generator to collapse: the tendency for a generator to focus on a subset of cover them. Experiments on standard mode collapse bench- modes to the exclusion of other parts of the target distribu- marks show that our method outperforms several competing tion [15]. Both empirical and theoretical studies have shown methods when addressing mode collapse. Our method also strong evidence that real data has a highly multi-modal dis- performs well on large-scale datasets such as ImageNet and tribution [41, 47]. Unconditional GANs trained on such data Places365, improving both image diversity and standard distributions often completely miss important modes, e.g., quality metrics, compared to previous methods. not being able to generate one of ten digits for MNIST [45], or omitting object classes such as people and cars within 1 synthesized scenes [5]. Class-conditional GANs alleviate fundamental challenges for GANs training [15, 45, 26]. this issue by enforcing labels that require the generator to Several solutions to mode collapse have been proposed, cover all semantic categories. However, in practice, it is including amending the adversarial loss to look ahead sev- often expensive to obtain labels for large-scale datasets. eral moves (e.g., Unrolled GAN [37]), jointly training an In this work, we present a simple but effective training encoder to recover latent distributions (e.g., VEEGAN [46]), method, self-conditioned GANs, to address mode collapse. adding an auxiliary loss to prevent the catastrophic forget- We train a class-conditional GAN and automatically obtain ting [9], packing the discriminator with sets of points instead image classes by clustering in the discriminator’s feature of singletons [45, 32] and training a mixture of generators space. Our algorithm alternates between learning a fea- with an auxiliary diversity objective [19, 13]. Different from ture representation for our clustering method and learning the above work, our method partitions the real distribution a better generative model that covers all the clusters. Such instead of the generated distribution, and devotes a class- partitioning automatically discovers modes the generator is conditioned discriminator to each target partition. currently missing, and explicitly requires the generator to Another related line of research trains class-conditional cover them. Figure1 shows several discovered clusters and GANs on unlabelled images by clustering on features ob- corresponding generated images for each cluster. tained via unsupervised feature learning methods [34, 44] Empirical experiments demonstrate that this approach or pre-trained classifiers [44]. In contrast, our method di- successfully recovers modes on standard mode collapse rectly clusters on discriminator features that inherently exist benchmarks (mixtures of Gaussians, stacked MNIST, in GANs, leading to a simpler method and achieving higher CIFAR-10). More importantly, our approach scales well quality generation in our experiments (Section 4.4). Mixture to large-scale image generation, achieving better Fréchet Density GAN proposes to use log-likelihoods of a Gaussian Inception Distance [1], Fréchet Segmentation Distance [5], mixture distribution in discriminator feature space as the and Inception Score [45] for both ImageNet and Places365, GAN objective [12]. GAN-Tree uses clustering to split a compared to previous unsupervised methods. GAN into a tree hierarchy of GANs for better mode cov- Our code and models are available on our website. erage [30]. These methods, while also using clustering or mixture models, are mostly orthogonal with our work. Fur- 2. Related Work thermore, the simplicity of our method allows it to be easily combined with a variety of these techniques. Generative Adversarial Networks (GANs). Since the in- troduction of GANs [16], many variants have been pro- 3. Method posed [38, 11, 43, 45, 2, 35, 17, 36], improving both the training stability and image quality. Due to its rapid advance, One of the core problems in generating diverse outputs in GANs have been used in a wide range of computer vision a high-dimensional space such as images is mode collapse: and graphics applications [48, 23, 49, 55, 21, 20]. GANs the support of the generator’s output distribution can be excel at synthesizing photorealistic images for a specific much smaller than the support of the real data distribution. class of images such as faces and cars [24, 25]. However, for One way mode collapse has been empirically lessened is by more complex datasets such as ImageNet, state-of-the-art use of a class-conditional GAN, which explicitly penalizes models are class-conditional GANs that require ground truth the generator for not having support on each class. image class labels during training [7]. To reduce the cost We propose to exploit this class-conditional architecture, of manual annotation, a recent work [34] presents a semi- but instead of assuming access to true class labels, we will supervised method based on RotNet [14], a self-supervised synthesize labels in an unsupervised way. On a high level, image rotation feature learning method. The model is trained our method dynamically partitions the real data space into with labels provided on only a subset of images. On the con- different clusters, which are used to train a class-conditional trary, our general-purpose method is not image-specific, and GAN. Because generation conditioned on a cluster index fully unsupervised. Section 4.4 shows that our method out- is optimized with respect to the corresponding conditioned performs a RotNet-based baseline. A recent method [40] discriminator, and each discriminator is responsible for a dif- proposes to obtain good clustering using GANs, while we ferent subset of the real distribution, our method encourages aim to achieve realistic and diverse generation. the generator output distribution to cover all partitions of the real data distribution. Mode collapse. Although GANs are formulated as a min- Specifically, to train a GAN to imitate a target distribution imax game in which each generator is evaluated against preal, we partition the data set into k clusters fπ1; : : : ; πkg a discriminator, during optimization, the generator faces a that are determined during training. No ground-truth la- slowly-changing discriminator that can guide generation to bels are used; training samples are initially clustered in the collapse to the point that maximizes the discriminator [37]. randomly initialized discriminator feature space, and the Mode collapse does occur in practice, and it is one of the clusters are updated periodically. A class-conditional GAN cluster data using discriminator features Algorithm 1 Self-Conditioned GAN Training (a) Initialize generator G and discriminator D Partition dataset into k sets fπ1; :::; πkg using Df outputs training for number of training epochs do π D real data c // Conditional GAN training based on current partition for number of training iterations for an epoch do for j in f1; :::; mg do (j) c Sample cluster c ∼ Pπ, where c is chosen with probability proportional to jπcj. (b) (j) (j) Sample image x ∼ πc(j) from cluster c . Sample latent z(j) ∼ N (0;I). z G D fake end for c Update G and D according to minG maxD LGAN (j) (j) (j) on minibatch f(c ; x ; z )gj. Eqn. (3) end for // Clustering to obtain new partitions Figure 2: We learn a discriminator D and a generator G that are Cluster on Df outputs of a subset of training set to new both conditioned on the automatically discovered cluster c. (a) identify a new partition fπc g into k sets, using For a specific c, the discriminator D must learn to recognize real previous centroids as initialization. Eqn. (2) new images sampled from the cluster πc of the dataset, and distinguish Find the matching ρ(·) between fπc gc and fπcgc them from (b) fake images synthesized by the class-conditional that minimizes Lmatch. Eqn. (3) generator G. (b) The class-conditional generator G synthesizes new Update all πc πρ(c). images from z. By learning to fool D when also conditioned on end for c, the generator learns

View Full Text

Details

  • File Type
    pdf
  • Upload Time
    -
  • Content Languages
    English
  • Upload User
    Anonymous/Not logged-in
  • File Pages
    17 Page
  • File Size
    -

Download

Channel Download Status
Express Download Enable

Copyright

We respect the copyrights and intellectual property rights of all users. All uploaded documents are either original works of the uploader or authorized works of the rightful owners.

  • Not to be reproduced or distributed without explicit permission.
  • Not used for commercial purposes outside of approved use cases.
  • Not used to infringe on the rights of the original creators.
  • If you believe any content infringes your copyright, please contact us immediately.

Support

For help with questions, suggestions, or problems, please contact us