Generative Adverserial Networks
Juan Vera
March 2025
Abstract
just a whiteboard, dm me if errors
Preliminaries.
A new framework for estimating generative models by training a generative network that captures the distribution of the data and a discriminative network which classifies whether the output of the generator is to have come from the training data itself.
The goal of the generator is to maximize the probability of the discriminator making a mistake, essentially generator and discriminator play a minimax game where one attempts to maximize for the objective one is trained to minimize.
Adverserial Nets.
The generator takes in a random noise variable .
The discriminator takes in , a vector of data (real or synthetic). is trained to maximize the probability of the output of being drawn / generated from the training data .
is trained to minimze such that the outputs of classify that correctly generated synthetic data from the real data.
The entire objective function is constructed as:
Note that
where we want to find the which minimizes the value function but find the which maximizes .
We're essentially playing a mini-max game, where is trained to maximize the value , which essentially translates to being an optimal binary classifier between synthetic data non-synthetic data.
This is as if , then , then , and finally
When we attempt to find which satisfies , we're try to find the which turns such that
Reminder that is assumed to be a clasifier where indicates that the input is real data while indicates that the input is fake synthetic data
Then as we try to find the which satisfies while we try to find the which satisfies , we're essentially playing what's called a zero-sum game where the gain in the generator is equal to the loss in the discrmininator (or vice versa).
Training.
For epochs and iterations = count of minibatches, update the discriminator by updating it's parameters, via gradient ascent,
while you update the generator by updating it's parameters via gradient descent,
all via the update rule,
where and if , else if .
The GAN is optimally trained when the discriminator can't distinguish the real data from the synthetic data -- this happens as we're trianing via gradient ascent which "confuses" or "fools" the discriminator from being able to optimally distinguish from both.