Deep Convolutional Generative Adversial Networks (DCGAN)

# Deep Convolutional Generative Adversial Networks (DCGAN)

Licence : Creative Commons Attribution 4.0 International (CC BY-NC-SA 4.0)
Last revision : June 14, 2021 09:35
Lectures project page: https://github.com/jeremyfix/deeplearning-lectures/

## Objectives

In this labwork, we aim at experimenting with generative networks and in particular the recently introduced Generative Adversial Networks (Goodfellow et al., 2014). Although other neural network architectures exist for learning to generate synthetic data from real observations (see for example this OpenAI blogpost which mentions some), the recently introduced GANs framework has shown to be efficient for generating a wide variety of data.

A GAN network is built from actually two networks that play a two player game :

• a generator which tries to generate images as real as possible, hopefully fooling the second player,
• a critic which tries to distinguish the real images from the fake images

The loss used for training these two neural networks reflect the objective of the generator to fool the critic and of the critic to correctly separate the real from the fake.

The generator generates an image from a random seed, $$z$$, say drawn from a normal distribution $$\mathcal{N}(0, 1)$$. Let us denote $$\mathcal{G}(z)$$ the output image (for now, we slightly postpone the discussion about the architecture used to generate an image). Let us denote by $$\mathcal{D}(x) \in [0, 1]$$ the score assigned by the critic to an image where $$\mathcal{D}(x) \approx 1$$ if $$x$$ is real and $$\mathcal{D}(x) \approx 0$$ if $$x$$ is a fake. The critic solves a binary classification problem with a binary cross entropy loss and seeks to minimize :

$\mathcal{L}_c = \frac{1}{2m} \sum_{i=1}^m -\log(D(x_i)) - \log(1-D(G(z_i)))$

You may recognize the usual binary cross entropy loss where the labels of the $$m$$ real data is set to $$y^x_i=1$$ and the labels of the $$m$$ fake data is set to $$y^z_i=0$$. This loss is to be minimized with respect to the parameters of the critic $$\theta_c$$.

The generator on his side wants to fool the critic and therefore wants its samples to be considered as real data. Therefore, it seeks to minimize:

$\mathcal{L}_g = \frac{1}{m} \sum_{i=1}^m -\log(D(G(z_i)))$

This loss is to be minimised with respect to the parameters of the generator $$\theta_g$$. The original paper considered fully connected architectures for the critic and the generator but the later work of Deep Convolutional GAN (DCGAN, (Radford, Metz, & Chintala, 2016)) proposed to use convolutional networks for the discriminator and the generator.

Our aim is to generate fake data and you are free to choose between several datasets : MNIST, FashionMNIST, EMNIST or even the colored house numbers SVHN. The image below show you some examples of digits generated by the generator you will train in this labwork on MNIST and on SVHN

## Lab work materials

You are provided with some starter code which already implements some functionalities :

• models.py : code defining the GAN
• main.py : code containing the training and testing functions
• mnist_generator.pt : a pretrained generator used for the end of the labwork pretrained on MNIST
• svhn_generator.pt : a pretrained generator used for the end of the labwork pretrained on SVHN

You will also need to install an external python dependency, the deepcs package, that you can install with :

mymachine:~:mylogin$python3 -m pip install --user deepcs  ## Implementing the Vanilla GAN ### Implementing the critic The critic is a simple convolutional neural network which has to stay if the input image is a real or a fake. You are free to experiment with any architecture but I can suggest you one. Denote by CBlock(k) the following sequence of layers : • $$2 \times [$$Conv($$k$$ channels, 3x3, stride=1, padding=same) - BatchNorm - LeakyRelu(0.2)$$]$$ • Conv($$k$$ channels, 3x3, stride=2, padding=same) - LeakyRelu(0.2) • Dropout(0.3) The architecture for the discriminator I propose you is : • CBlock(32) - CBlock(64) - CBlock(96) - Linear(1) Every block is downsampling the representation so that with a $$(B, 1, 28, 28)$$ input, we end the convolutional part with a $$(B, 96, 4, 4)$$ and the linear layer has therefore $$1536$$ weights and $$1$$ bias. The output of the network is the logit, i.e. before the application of the sigmoid which is actually embedded in the BCEWithLogitsLoss we will be using. Exercice Implement the critic in the models.py script. You have to define the neural network in the constructor Discriminator class and implement the forward method. Note that since every convolutional layer is followed by a batch-normalization, you can remove the bias from the convolutional layer that would anyway be canceled by the normalization (see the constructor of Conv2d). Sanity check : def test_discriminator(): critic = Discriminator((1, 28, 28), 0.3, 32) X = torch.randn(64, 1, 28, 28) out = critic(X) assert(out.shape == torch.Size([64]))  ### Implementing the generator The generator takes as input a $$(B, N_z)$$ normally distributed tensor and has to produce a $$28\times 28$$ grayscale image. While the DCGAN paper suggested to use fractionnally strided convolutions (or transposed convolutions), this can introduce artifacts if not properly tuned. We rather consider the alternative proposed by (Odena, Dumoulin, & Olah, 2016) which is to perform a bilinear upsampling followed by a convolution. Let us denote by GBlock(k) the following sequence of layers: • UpSample(scale_factor=2) • Conv(k channels, 3x3, padding=same) - BatchNorm - ReLU • Conv(k channels, 3x3, padding=same) - BatchNorm - ReLU The proposed architecture of the generator is: • Linear($$7\times7\times256$$) - BatchNorm - ReLU • GBlock(128) • GBlock(64) • Conv(1 channel, 1x1) - Tanh If you use SVHN, you should start with a Linear($$8\times8\times256$$) since SVHN images are $$32\times32$$ while the MNIST like data are $$28\times28$$. In between the linear layer and the first convolution, note you will have to “reshape” the tensor (using the Tensor.view) method. The tanh activation for the last layer is suggested in (Radford et al., 2016) to be a good idea. Exercice Implement the generator in the models.py script. You have to create the network in the constructor of the Generator class and to implement the forward function. Note you can use the up_conv_bn_relu builder function provided in this script. The forward(X, batch_size) either takes as input a random vector $$X$$ or the number of samples you want. As for the critic, the bias is useless in the convolutional layers that are followed by a batch-normalization. Sanity check : def test_generator(): # Testing the generator for producing MNIST like data # Note: if you use SVHN, change these to (3, 32, 32) # and adapt below generator = Generator((1, 28, 28), 100, 64) X = torch.randn(64, 100) out = generator(X, None) assert(out.shape == torch.Size([64, 1, 28, 28])) out = generator(None, 64) assert(out.shape == torch.Size([64, 1, 28, 28]))  Note that the generator is outputting values in $$[-1, 1]$$. You may also notice in the dataloaders that the real images are rescaled in $$[-1, 1]$$ to guarantee that both the real and fake images lie in the same range of values. ### Implementing the GAN The GAN network is the discriminator along with the generator. Exercice Fill in the missing code in the forward method of the GAN class. Note that the forward method has two modes, either accepting a tensor (that we expect are real images) or a number of images to sample. Its output is always the pair of logits and images. ### Implementing the optimizers, losses and backprop It is now time to implement the losses in the main.py script. The critic and the generator are trained separately, one after the other. Both training involves the binary cross entropy loss on the logits output of the critic : • the discriminator wants to classify the real images as positive and the fake as negatives • the generator wants its fake images to be classified as positive by the discriminator When we train the discriminator, only the parameters of the discriminator are expected to be modified. When we train the generator (even though its loss goes through the discriminator network), only the parameters of the generators are expected to be modified. Therefore, two optimizers will be defined : one for the parameters of the critic, one for the parameters of the generator. For the loss, we need a binary cross entropy loss taking as input the logits Exercice Define the optimizers and loss in the main.py script. The next step is to compute the backward propagation on the right losses. If we denote $$r_i$$ and $$f_i$$ the logits assigned by the critic to respectively a minibatch $$X$$ of real images and a minibatch of size $$b_i$$ of fake images, the loss to be minimized by the critic is: $\mathcal{L}_c = \frac{1}{2m} \sum_{i=1}^m -\log(D(x_i)) - \log(1-D(G(z_i)))$ $\mathcal{L}_c = BCELoss(\begin{bmatrix}r_0 \\ \vdots \\ r_{b_i-1} \\ f_0 \\ \vdots \\ f_{b_i - 1}\end{bmatrix}, \begin{bmatrix} 1 \\ \vdots \\ 1 \\ 0 \\ \vdots \\ 0 \end{bmatrix}) = -\frac{1}{2b_i} (\sum_i \log(r_i) + \sum_k \log(1-f_i)) = \frac{BCELoss(\begin{bmatrix}r_0 \\ \vdots \\ r_{b_i-1} \end{bmatrix}, \begin{bmatrix}1 \\ \vdots \\ 1\end{bmatrix}) + BCELoss(\begin{bmatrix} f_0 \\ \vdots \\ f_{b_i - 1}\end{bmatrix}, \begin{bmatrix}0 \\ \vdots \\ 0\end{bmatrix}) }{2}$ Exercice Implement the above loss in the main.py script within the training loop. Note that you have two vectors named pos_labels and neg_labels containing respectively only ones and zeros. Implement also the three lines for the backward pass (reset the gradient accumulator, perform the backward pass, update the parameters). Finally, for the generator, since it wants to fool the critic, denoting $$f_i$$ the logits assigned by the critic to a newly sampled set of fake images, the loss to be minimized by the generator is : $\mathcal{L}_g = BCELoss(\begin{bmatrix} f_0 \\ \vdots \\ f_{b_i - 1}\end{bmatrix}, \begin{bmatrix} 1 \\ \vdots \\ 1 \end{bmatrix}) = -\frac{1}{b_i} (\sum_i \log(f_i))$ Exercice Implement the above loss in the main.py script within the training loop. Note that you have the vector named pos_labels containing respectively only ones. Implement also the three lines for the backward pass (reset the gradient accumulator, perform the backward pass, update the parameters) for the generator. ### Training You can now start training your networks by running the main script. The default arguments given to argparse should be reasonably good. Do not forget to start the tensorboard and to look at it, at every epochs, some generated samples are written on it. During the first 5 epochs, you should already observe some kind of written digits. Your model should have almost 2M parameters and it takes 30s. per epoch. By default, the dataset used is MNIST but you can change this to the dataset you want (check the –dataset option of the main.py script). For example, for training on SVHN : mymachine:~:mylogin$ python3 main.py train --dataset SVHN


Below is an example of digits generated during a successful training with a noise vector defined once for all before training (i.e. we always plot the generated image associated with the same random inputs) on MNIST.

While training, you can move on the next section where you will load a pretrained network.

### Generating fake images

To generate new samples, we just need to evaluate the generator of normally distributed inputs. For every random input vector, you get a fake image that looks hopefully realistic.

Exercice Download the pretrained generator generator.pt that has been pretrained for 400 epochs (which is by far more than necessary). Fill in the code in the generate function of the main.py script and run some generation of fake images by issuing :

mymachine:~:mylogin\$ python3 main.py generate --modelpath generator.pt


Note that the generated images have been already denormalized in the code you are provided.

### Interpolating in the latent space

There is an interesting property of the generator which is that if you move in the latent space you continuously move in the digits space.

Exercice Generate two random vectors $$z_1, z_2$$ drawn from $$\mathcal{N}(0, 1)$$, compute their image through the generator $$G(z_1), G(z_2)$$ as well as the image of the linearly interpolated noise vectors $$G(z_1 + \alpha (z_2 - z_1)), \alpha \in [0, 1]$$.

## Going further

At the time of writing the subject, I’m quite surprised by the values of the losses and accuracies. Indeed, the generator seems to be outputting realistic images but the discriminator accuracy converges up to 96$$\%$$ which is really surprising since we may expect the discriminator to fail distinguishing the real from the fake. Apparently it succeeds to differentiate both even thought visually, the digits seems pretty realistic. As I understand it, it does not seem to be related to mode collapse because interpolating in the latent space shows a large diversity of samples. The generator loss is also unexpectedly staying high. This is surprising given the apparently realistic outputs produced by the generator but, at least, this is in agreement with the loss/accuracy of the discriminator which fails to be fooled by the generator.

GAN can have problems being trained. Sometimes, training is unstable. Sometimes, training leads to mode collapse, a situation where the generator fails to produce diversity in its output and the model gets locked in this mode. Variations, known as Wasserstein GAN (WGAN, (Arjovsky, Chintala, & Bottou, 2017)), Wasserstein GAN with gradient penalty (Gulrajani, Ahmed, Arjovsky, Dumoulin, & Courville, 2017) and others were introduced to fix these issues. However, a recent paper (Lucic, Kurach, Michalski, Gelly, & Bousquet, 2018) suggests that with enough hyperparameters tuning, “even” the vanilla GAN can work as well as its variations.

Finally, for evaluating the quality of the generated image, the literature is currently on the Fréchet Inception Distance introduced in (Heusel, Ramsauer, Unterthiner, Nessler, & Hochreiter, 2018).

## References

Arjovsky, M., Chintala, S., & Bottou, L. (2017). Wasserstein generative adversarial networks. In D. Precup & Y. W. Teh (Eds.), Proceedings of the 34th international conference on machine learning (Vol. 70, pp. 214–223). International Convention Centre, Sydney, Australia: PMLR. Retrieved from http://proceedings.mlr.press/v70/arjovsky17a.html

Goodfellow, I., Pouget-Abadie, J., Mirza, M., Xu, B., Warde-Farley, D., Ozair, S., … Bengio, Y. (2014). Generative adversarial nets. In Z. Ghahramani, M. Welling, C. Cortes, N. Lawrence, & K. Q. Weinberger (Eds.), Advances in neural information processing systems (Vol. 27, pp. 2672–2680). Curran Associates, Inc. Retrieved from https://proceedings.neurips.cc/paper/2014/file/5ca3e9b122f61f8f06494c97b1afccf3-Paper.pdf

Gulrajani, I., Ahmed, F., Arjovsky, M., Dumoulin, V., & Courville, A. (2017). Improved training of wasserstein gans. Retrieved from http://arxiv.org/abs/1704.00028

Heusel, M., Ramsauer, H., Unterthiner, T., Nessler, B., & Hochreiter, S. (2018). GANs trained by a two time-scale update rule converge to a local nash equilibrium. Retrieved from http://arxiv.org/abs/1706.08500

Lucic, M., Kurach, K., Michalski, M., Gelly, S., & Bousquet, O. (2018). Are gans created equal? A large-scale study. In S. Bengio, H. Wallach, H. Larochelle, K. Grauman, N. Cesa-Bianchi, & R. Garnett (Eds.), Advances in neural information processing systems (Vol. 31, pp. 700–709). Curran Associates, Inc. Retrieved from https://proceedings.neurips.cc/paper/2018/file/e46de7e1bcaaced9a54f1e9d0d2f800d-Paper.pdf

Odena, A., Dumoulin, V., & Olah, C. (2016). Deconvolution and checkerboard artifacts. Distill. https://doi.org/10.23915/distill.00003