Generative Adversarial Networks (GANs) are a type of AI that can create new, synthetic data that's similar to real data. This is done by pitting two neural networks against each other in a game of cat and mouse.
The generator network creates fake data, while the discriminator network tries to distinguish between real and fake data. This process is repeated until the generator creates data that's indistinguishable from real data.
GANs have many potential applications, including image and video generation, data augmentation, and even art. They've been used to generate realistic images of faces, objects, and even entire scenes.
One of the key benefits of GANs is their ability to learn from a small dataset, making them particularly useful for tasks where data is scarce.
What Is
A Generative Adversarial Network (GAN) is a type of neural network used for unsupervised learning. This means it can create new data without being explicitly told what to generate.
GANs are made up of two neural networks: a Generator and a Discriminator. The Generator attempts to produce realistic data, while the Discriminator tries to distinguish between real and fake data.
The training process for GANs is a competition between the Generator and the Discriminator. The Generator tries to create better and more realistic data, while the Discriminator tries to differentiate between real and fake data.
GANs can be broken down into three parts: Generative, Adversarial, and Networks. The Generative part learns a generative model, which describes how data is generated in terms of a probabilistic model. The Adversarial part involves setting one thing up against another, such as comparing the generative result with the actual images in the data set. The Networks part uses deep neural networks as artificial intelligence (AI) algorithms for training purposes.
Here are some key characteristics of GANs:
- GANs are a type of generative model based on deep learning.
- They consist of a Generator that produces data and a Discriminator that evaluates its realism.
- The Generator takes a random noise vector as input and generates new data based on that input.
- The Discriminator has two output values – 1 or 0 – depending on whether the data is real or fake.
GANs have been used in various applications, including image synthesis, style transfer, and text-to-image synthesis. They have also revolutionized generative modeling.
Mathematical Background
The original GAN is defined as a game between a generator and a discriminator, where the generator's strategy set is the set of all probability measures on the probability space (Ω, μref).
The generator's goal is to match its output distribution to the reference distribution μref as closely as possible.
The generator's task is to approach μG ≈ μref, that is, to match its own output distribution as closely as possible to the reference distribution.
The discriminator's strategy set is the set of Markov kernels μD:Ω → P[0,1], where P[0,1] is the set of probability measures on [0,1].
The discriminator's task is to output a value close to 1 when the input appears to be from the reference distribution, and to output a value close to 0 when the input looks like it came from the generator distribution.
The GAN game is a zero-sum game, with the objective function L(μG, μD) defined as the expectation of the logarithm of the discriminator's output for inputs from both the reference and generator distributions.
The generator aims to minimize the objective function, while the discriminator aims to maximize it.
See what others are reading: Pre-trained Multi Task Generative Ai Models Are Called
Training and Evaluation
Training a Generative Adversarial Network (GAN) involves training the generator and discriminator simultaneously in a minimax game. The generator tries to create realistic samples to fool the discriminator, while the discriminator tries to classify real and generated samples correctly.
The evaluation of a GAN can be challenging, but common methods include using a combination of qualitative and quantitative measures. Qualitative measures involve visually inspecting and comparing the generated samples to real ones, while quantitative measures involve using metrics such as the Inception score, Frechet Inception Distance (FID), and Wasserstein distance to evaluate the generated samples' quality numerically.
To train a GAN, you can use a custom training loop, specifying options such as the mini-batch size, number of epochs, and Adam optimization parameters. It's also important to add noise to the real data by randomly flipping the labels assigned to the real images to balance the learning of the discriminator and the generator.
Intriguing read: Learning Generative Ai
Training and Evaluating
Training a Generative Adversarial Network (GAN) is a complex process that requires careful consideration of several factors. The two main components of a GAN, the generator and discriminator, are trained simultaneously in a minimax game where the generator tries to create realistic samples to fool the discriminator, while the discriminator tries to classify real and generated samples correctly.
Training the discriminator involves using both real and fake data to teach the binary classifier to distinguish between the two. The discriminator's goal is to correctly identify whether the input is real or fake, and its loss function is typically calculated using binary cross-entropy.
The generator, on the other hand, is trained to produce synthetic data that can deceive the discriminator. Its loss function is typically calculated using a combination of the generator's ability to produce realistic samples and the discriminator's ability to correctly identify them.
During training, the generator and discriminator are updated simultaneously in an adversarial process. The generator generates fake data, which is then fed to the discriminator, and the discriminator's loss is calculated based on its ability to correctly identify the fake data. The generator's loss is then calculated based on the discriminator's loss, and the generator's weights are updated to try to minimize its loss.
See what others are reading: Telltale Words Identify Generative Ai Text
To monitor the training progress, it's essential to track the loss functions of both the generator and discriminator. This can provide insights into the training process and help identify potential problems, such as mode collapse or poor convergence.
Some common metrics used to evaluate the performance of a GAN include the Inception score, Frechet Inception Distance (FID), and Wasserstein distance. These metrics can provide a quantitative measure of the quality of the generated samples and help identify areas for improvement.
Here are some common metrics used to evaluate the performance of a GAN:
By carefully monitoring the training process and using a combination of qualitative and quantitative metrics, you can train a GAN that produces high-quality, realistic samples.
Loading the Dataset
Loading the Dataset is a crucial step in the training and evaluation process. A CIFAR-10 dataset is created for training with a specific code that specifies a root directory.
This code also turns on train mode and downloads the dataset if needed. It applies the specified transform to the data. Subsequently, it generates a 32-batch DataLoader.
This DataLoader shuffles the training set of data. This is essential for preventing the model from learning patterns based on the order of the data.
For your interest: Generative Ai Code
Vanishing Gradient
The vanishing gradient problem is a significant challenge in training generative adversarial networks (GANs). It occurs when the generator has access to a restricted strategy set, which takes up a vanishingly small proportion of its entire strategy set.
In practice, this means the generator only has access to measures of form μ μ Z∘ ∘ Gθ θ − − 1{\displaystyle \mu _{Z}\circ G_{\theta }^{-1}}, where Gθ θ {\displaystyle G_{\theta }} is a function computed by a neural network with parameters θ θ {\displaystyle \theta }, and μ μ Z{\displaystyle \mu _{Z}} is an easily sampled distribution.
The discriminator also has a restricted strategy set, which can lead to unstable convergence. This is because the equilibrium point is no longer guaranteed when they have a restricted strategy set.
To improve convergence stability, some training strategies start with an easier task and gradually increase the difficulty during training, essentially applying a curriculum learning scheme.
You might enjoy: When Was Generative Ai Open Source
Variants and Applications
Generative Adversarial Networks (GANs) have been used to up-scale low-resolution 2D textures in old video games by recreating them in 4k or higher resolutions via image training.
Some prominent GAN variants include those that can generate high-resolution images from low-resolution inputs, allowing for improved image quality in various applications.
GANs have been employed in a wide range of applications, including image generation, data augmentation, style transfer, super-resolution, and image-to-image translation.
GANs can modify the style of images, such as converting photos into the style of famous paintings or changing day scenes to night scenes.
Here are some of the widely recognized uses of GANs:
- Image Synthesis and Generation
- Image-to-Image Translation
- Text-to-Image Synthesis
- Data Augmentation
- Data Generation for Training
Variants
There is a veritable zoo of GAN variants.
Some of the most prominent ones include the Conditional GAN, which is used for tasks like image-to-image translation.
The Generative Adversarial Network (GAN) has many different applications, but its variants are what make it so versatile.
A notable variant is the Deep Convolutional GAN (DCGAN), which is particularly useful for generating high-quality images.
GAN variants have been used in a wide range of fields, from art to medicine, and have shown great promise in many areas.
The StyleGAN is another variant that has gained popularity, especially in the field of computer vision.
A unique perspective: How Multimodal Used in Generative Ai
Conditional
Conditional GANs are a type of Generative Adversarial Network that allows the generator network to be conditioned on an additional input vector.
This conditioning can generate images corresponding to a specific class, category, or attribute. The architecture of a cGAN is similar to that of a traditional GAN, with a condition vector concatenated with the input noise vector fed into the generator.
The discriminator network is also modified to take in both the generated image and the condition vector as inputs. This allows for greater control over the generation process, as the generator can be conditioned on specific attributes or classes.
For example, a conditional GAN can be used to generate images of a specific class, such as cats or dogs. In 2017, a conditional GAN learned to generate 1000 image classes of ImageNet.
Conditional GANs are particularly useful in applications such as image editing or synthesis, where the goal is to generate images with specific properties or characteristics. This can be achieved by conditioning the generator on specific attributes or classes.
Suggestion: Can I Generate Code Using Generative Ai Models
Some examples of applications of cGANs include image-to-image translation, where the goal is to generate images of one domain that match the style or content of another domain, and text-to-image synthesis, where the goal is to generate images based on textual descriptions.
Here are some key advantages of cGANs:
- Greater control over the generation process
- Ability to generate images with specific properties or characteristics
- Useful in applications such as image editing or synthesis
- Can be used for image-to-image translation and text-to-image synthesis
Autoencoder
An autoencoder is a type of neural network that can be used for dimensionality reduction and generative modeling.
The idea behind an autoencoder is to train a neural network to learn a compressed representation of the input data, often referred to as a latent vector.
An autoencoder typically consists of an encoder and a decoder, where the encoder maps the input data to a lower-dimensional space, and the decoder maps the latent vector back to the original input data.
Autoencoders can be trained to reconstruct the input data with high accuracy, making them useful for tasks such as anomaly detection and data denoising.
The adversarial autoencoder (AAE) is a variant of the autoencoder that incorporates a discriminator to improve the quality of the generated data.
Suggestion: Vector Database Generative Ai
Utility Class
A utility class is a crucial component in building a Generative Adversarial Network (GAN). It helps to define the architecture of the discriminator, which is a critical component of the GAN.
In PyTorch, a utility class is used to define the discriminator architecture. The class Discriminator is descended from nn.Module, making it a PyTorch module.
The discriminator architecture is composed of linear layers, batch normalization, dropout, convolutional, LeakyReLU, and sequential layers. This architecture is used to determine the validity of an image, i.e., whether it is real or artificial.
The output of the discriminator is a probability value that indicates the likelihood of the input image being real. This probability value is a key component in training the GAN.
A utility class can be used to define different types of discriminators, such as a convolutional discriminator or a linear discriminator. Each type of discriminator has its own architecture and is used for specific tasks.
For your interest: Generative Ai Architecture Diagram
Here is an example of how a utility class can be used to define a discriminator:
```python
class Discriminator(nn.Module):
def __init__(self):
super(Discriminator, self).__init__()
self.conv1 = nn.Conv2d(1, 64, kernel_size=3)
self.conv2 = nn.Conv2d(64, 128, kernel_size=3)
self.conv3 = nn.Conv2d(128, 256, kernel_size=3)
self.fc1 = nn.Linear(256*3*3, 128)
self.fc2 = nn.Linear(128, 1)
self.leaky_relu = nn.LeakyReLU(0.2)
self.dropout = nn.Dropout(0.3)
def forward(self, img):
img = self.leaky_relu(self.conv1(img))
img = self.leaky_relu(self.conv2(img))
img = self.leaky_relu(self.conv3(img))
img = img.view(-1, 256*3*3)
img = self.leaky_relu(self.fc1(img))
img = self.dropout(img)
img = self.fc2(img)
return img
```
This utility class can be used to define different types of discriminators by changing the architecture of the class.
Explore further: Generative Ai and Cybersecurity
Comparison of Large Language Models
Large Language Models (LLMs) have emerged as a cornerstone in the advancement of artificial intelligence, transforming our interaction with technology and our ability to process and generate human language.
LLMs have been instrumental in improving language understanding and generation capabilities, allowing for more accurate and context-specific responses. This has far-reaching implications for various industries and applications.
The comparison of LLMs highlights their varying strengths and weaknesses, with some exceling in specific tasks such as question-answering or text summarization.
A fresh viewpoint: Generative Ai and Llms for Dummies
Implementation and Usage
The implementation of a Generative Adversarial Network (GAN) involves a series of steps that can be tricky to follow, but don't worry, we've got you covered.
The code for a DC-GAN architecture with a generator and discriminator is remarkably simple, implementing a minimax game where the generator creates realistic-looking images and the discriminator tries to distinguish between real and fake images.
Here are the key components of the generator model: dense layers, transposed convolutional layers, batch normalization, and a final layer with a 'tanh' activation function.
The discriminator model consists of convolutional layers followed by dense layers, processing the input images and gradually down-sampling them using LeakyReLU activations and batch normalization.
The loss functions are designed to penalize the discriminator for misclassifying real images and for being fooled by the generator, while the generator is penalized if the discriminator can easily tell its images are fake.
Here are the main functions involved in the training loop:
- train_step: generates fake images, evaluates them with the discriminator, and updates the generator and discriminator parameters.
- train: calls the train_step function for each batch of real images in the dataset, for a specified number of epochs.
- generate_and_save_images: generates images from random noise, reshapes and scales them to visual range, and saves and displays them using Matplotlib.
Implementation of
Implementation of InfoGAN and DC-GAN involves understanding the architecture and components of each model.
The InfoGAN game is defined as follows: three probability spaces define an InfoGAN game, including the space of reference images, the fixed random noise generator, and the fixed random information generator.
A DC-GAN architecture with a generator and discriminator can be implemented to generate fake images resembling the MNIST dataset digits. The generator learns to create realistic-looking images, while the discriminator learns to distinguish between real and fake images.
The generator model consists of a series of dense layers and transposed convolutional layers, starting with a dense layer that reshapes the noise input into a 7x7 image with 256 channels.
The discriminator model consists of convolutional layers followed by dense layers, processing the input images and gradually down-sampling them using convolutional layers with LeakyReLU activations and batch normalization.
The objective function of the InfoGAN game is L(G,Q,D) = LGAN(G,D) - λI^ (G,Q), where LGAN(G,D) is the original GAN game objective, and I^ (G,Q) is the mutual information between the informative label part c and the generator's output G(z,c).
You might enjoy: Velocity Model Prediciton Using Generative Ai
The DC-GAN's training loop involves calling the train_step function for each batch of real images in the dataset, for a specified number of epochs. The train_step function generates random noise, passes it through the generator to produce fake images, and updates the generator and discriminator parameters using the Adam optimizer.
Here is a summary of the components involved in the DC-GAN architecture:
Importing Libraries
To get started with implementing a Generative Adversarial Network (GAN) in PyTorch, you'll need to import the required libraries. This includes importing torch, torch.nn, torch.optim, and torchvision.
The torch library is the foundation of PyTorch, providing a dynamic computation graph and automatic differentiation.
For the CIFAR-10 image dataset, you'll also need to import torchvision.datasets, torchvision.transforms, and torch.utils.data.
You might enjoy: Transfer Learning Pytorch
Parameters for Later Use
In the implementation of a Generative Adversarial Network (GAN), it's essential to define parameters that will be used in later processes.
The latent space's dimensionality is represented by latent_dim, which determines the complexity of the generated data.
To optimize the GAN's performance, we need to specify the learning rate of the optimizer, denoted as lr.
The Adam optimizer's coefficients, beta1 and beta2, also play a crucial role in the optimization process.
The total number of training epochs is represented by num_epochs, which affects the overall training time and model performance.
Here's a summary of the key parameters to keep in mind:
A Utility Class
A Utility Class is a crucial part of building a GAN, and it's essential to define it correctly. The Generator class in PyTorch inherits from nn.Module and is comprised of a sequential model with various layers.
The Generator class is used to synthesize an image from a latent vector, which is the generator's output. The architecture of the Generator class uses a series of learned transformations to turn the initial random noise in the latent space into a meaningful image.
A key aspect of the Generator class is its ability to take a latent vector and produce an image. The class is designed to be flexible and can be modified to suit the specific needs of your project.
The Discriminator class, on the other hand, is used to determine the validity of an image. It is composed of linear layers, batch normalization, dropout, convolutional, LeakyReLU, and sequential layers. The Discriminator class is also inherited from nn.Module and is designed to be highly effective in distinguishing between real and artificial images.
Here are some common layers used in the Generator and Discriminator classes:
- Tanh layer: used for activation
- Linear layer: used for fully connected layers
- Convolutional layer: used for image processing
- Batch normalization layer: used for normalization
- Reshaping layer: used for image resizing
- Upsampling layer: used for image upsampling
- LeakyReLU layer: used for activation
By using a utility class, you can easily build and modify the Generator and Discriminator classes to suit your project's needs. This approach also helps to keep your code organized and maintainable.
Advantages and Disadvantages
GANs have some amazing advantages that make them super useful. They can generate new, synthetic data that resembles some known data distribution, which can be useful for data augmentation, anomaly detection, or creative applications.
One of the biggest advantages of GANs is their ability to produce high-quality, photorealistic results in image synthesis, video synthesis, music synthesis, and other tasks. This is especially useful in applications where realistic results are crucial.
Here are some of the key advantages of GANs:
- Synthetic data generation
- High-quality results
- Unsupervised learning
- Versatility
However, GANs also have some significant disadvantages that need to be considered. For example, they can be difficult to train, with the risk of instability, mode collapse, or failure to converge.
Advantages
GANs have several advantages that make them a powerful tool in various applications.
GANs can generate new, synthetic data that resembles some known data distribution, which can be useful for data augmentation, anomaly detection, or creative applications.
One of the key benefits of GANs is their ability to produce high-quality, photorealistic results in image synthesis, video synthesis, music synthesis, and other tasks.
GANs can also be trained without labeled data, making them suitable for unsupervised learning tasks, where labeled data is scarce or difficult to obtain.
GANs are versatile and can be applied to a wide range of tasks, including image synthesis, text-to-image synthesis, image-to-image translation, anomaly detection, data augmentation, and others.
Here are some of the key advantages of GANs:
- Synthetic data generation
- High-quality results
- Unsupervised learning
- Versatility
Disadvantages of
GANs can be finicky to train, with a risk of instability, mode collapse, or failure to converge. This can be frustrating for developers, especially when working with high-stakes applications.
Training GANs requires significant computational resources and can be slow to train, especially for high-resolution images or large datasets. This can be a major bottleneck in many projects.
Overfitting is another issue with GANs, where they produce synthetic data that's too similar to the training data, lacking diversity. This can lead to synthetic data that's not very useful.
GANs can reflect the biases and unfairness present in the training data, leading to discriminatory or biased synthetic data. This is a major concern, especially in applications where fairness and transparency are crucial.
Here are some of the key disadvantages of GANs:
- Training Instability
- Computational Cost
- Overfitting
- Bias and Fairness
- Interpretability and Accountability
Future of Generative Adversarial Networks
Generative Adversarial Networks (GANs) have already shown their potential in various applications, including image and video synthesis, natural language processing, and drug discovery. They're going to find even more exciting applications in the future.
GANs can generate realistic 3D models of objects and environments for use in virtual and augmented reality applications. This will result in more immersive and lifelike experiences in fields like gaming and architectural design.
Fashion and design will also benefit from GANs, which can create new designs and patterns for clothing, accessories, and home decor. This will facilitate more creative and efficient design processes, as well as personalized products based on individual preferences.
Explore further: Generative Design Ai
In healthcare, GANs can help improve disease diagnosis and treatment by analyzing and synthesizing medical images. They can also assist drug discovery and development by generating new molecules with desired properties.
Here are some potential future applications of GANs:
- Virtual reality and augmented reality
- Fashion and design
- Healthcare
- Robotics
- Marketing and advertising
- Art and music
As GANs continue to improve, their applications will undoubtedly expand, providing exciting new opportunities for data generation and other fields.
Frequently Asked Questions
Are GANs still relevant in 2024?
Yes, GANs remain highly relevant in 2024, with ongoing improvements in stability and applications across various industries. Their innovative capabilities continue to drive advancements in fields like image generation and data augmentation.
Does Gen AI use GANs?
GANs were the most popular generative AI algorithm until recently, but Gen AI may not rely solely on GANs due to advancements in diffusion-based models and transformers.
Why is GAN better than cnn?
GANs outperform CNNs in generating new data that follows a pattern, whereas CNNs excel at recognizing existing patterns. If you're looking to create realistic images or text, GANs might be the better choice.
Sources
- AI Melody Generation from Lyrics (github.com)
- "This Person Does Not Exist" (boingboing.net)
- "StyleGAN: Official TensorFlow Implementation" (github.com)
- "ARTificial Intelligence enters the History of Art" (issuewire.com)
- cs.CV (arxiv.org)
- 1612.07919 (arxiv.org)
- 10.1109/TNN.2008.2000204 (doi.org)
- "Noise-Contrastive Estimation" (mlr.press)
- 10.1145/2463372.2465801 (doi.org)
- 10.1016/j.neunet.2020.04.008 (doi.org)
- 1906.04493 (arxiv.org)
- 10.3390/en13010130 (doi.org)
- "Samsung's AI Lab Can Create Fake Video Footage From a Single Headshot" (adweek.com)
- "ASTOUNDING AI GUESSES WHAT YOU LOOK LIKE BASED ON YOUR VOICE" (futurism.com)
- 10.1145/3397271.3401129 (doi.org)
- 10.1016/j.scs.2019.101602 (doi.org)
- 1905.06464 (arxiv.org)
- 10.1080/23729333.2019.1615729 (doi.org)
- 1905.02200 (arxiv.org)
- 2016arXiv160902612V (harvard.edu)
- 1609.02612 (arxiv.org)
- "Generating Videos with Scene Dynamics" (columbia.edu)
- 1707.02392 (arxiv.org)
- "3D Generative Adversarial Network" (mit.edu)
- 1702.01983 (arxiv.org)
- 10.1145/3424116 (doi.org)
- 1908.05551 (arxiv.org)
- 2005.12126 (arxiv.org)
- "Nvidia's AI recreates Pac-Man from scratch just by watching it being played" (theverge.com)
- 1809.00219 (arxiv.org)
- "AI Art at Christie's Sells for $432,500" (nytimes.com)
- 10.3390/arts8010026 (doi.org)
- cs.AI (arxiv.org)
- 1706.07068 (arxiv.org)
- 10.1007/978-3-030-50426-7_36 (doi.org)
- "The Rise of AI Supermodels" (cdotrends.com)
- Generative image inpainting with contextual attention (thecvf.com)
- "A never-ending stream of AI art goes up for auction" (theverge.com)
- "The Defense Department has produced the first tools for catching deepfakes" (technologyreview.com)
- "California laws seek to crack down on deepfakes in politics and porn" (cnet.com)
- "'This Person Does Not Exist' Website Uses AI To Create Realistic Yet Horrifying Faces" (slashdot.org)
- 10.1016/j.asoc.2020.106165 (doi.org)
- 10.1007/s11760-023-02878-4 (doi.org)
- "A Molecule Designed By AI Exhibits "Druglike" Qualities" (wired.com)
- 10.1038/s41587-019-0224-x (doi.org)
- 10.1007/s00162-020-00518-y (doi.org)
- 2001.09618 (arxiv.org)
- 10.1016/j.proci.2022.07.244 (doi.org)
- "Investigation of the generalization capability of a generative adversarial network for large eddy simulation of turbulent premixed reacting flows" (sciencedirect.com)
- 2308.16015 (arxiv.org)
- "Influence of adversarial training on super-resolution turbulence reconstruction" (aps.org)
- 10.1088/1748-0221/14/11/P11028 (doi.org)
- 2019JInst..14P1028A (harvard.edu)
- 1909.04451 (arxiv.org)
- "Deep generative models for fast shower simulation in ATLAS" (cern.ch)
- 10.1007/s41781-018-0015-y (doi.org)
- 2018arXiv180500850M (harvard.edu)
- 1805.00850 (arxiv.org)
- 10.1007/s41781-018-0019-7 (doi.org)
- 1807.01954 (arxiv.org)
- 1712.10321 (arxiv.org)
- 1705.02355 (arxiv.org)
- 10.1007/s41781-017-0004-6 (doi.org)
- 1701.05927 (arxiv.org)
- 10.1186/s40668-019-0029-9 (doi.org)
- 1706.02390 (arxiv.org)
- "CosmoGAN: Training a neural network to study dark matter" (phys.org)
- "Researchers Train a Neural Network to Study Dark Matter" (rdmag.com)
- 1702.00403 (arxiv.org)
- cs.LG (arxiv.org)
- 2102.07097 (arxiv.org)
- "Alias-Free Generative Adversarial Networks (StyleGAN3)" (nvlabs.github.io)
- 10.1109/cvpr42600.2020.00813 (doi.org)
- 1912.04958 (arxiv.org)
- 10.1109/cvpr.2019.00453 (doi.org)
- 1812.04948 (arxiv.org)
- "A Style-Based Generator Architecture for Generative Adversarial Networks" (doi.org)
- 10.1109/iccv.2019.00467 (doi.org)
- 1905.01164 (arxiv.org)
- "SinGAN: Learning a Generative Model from a Single Natural Image" (doi.org)
- "Training Generative Adversarial Networks with Limited Data" (neurips.cc)
- 2006.10738 (arxiv.org)
- "Differentiable Augmentation for Data-Efficient GAN Training" (neurips.cc)
- "A Gentle Introduction to BigGAN the Big Generative Adversarial Network" (machinelearningmastery.com)
- 1611.07004 (arxiv.org)
- 1703.10593 (arxiv.org)
- "Bidirectional Generative Adversarial Networks for Neural Machine Translation" (aclweb.org)
- 1606.03657 (arxiv.org)
- stat.ML (arxiv.org)
- 1606.00704 (arxiv.org)
- 1605.09782 (arxiv.org)
- "InfoGAN: Interpretable Representation Learning by Information Maximizing Generative Adversarial Nets" (neurips.cc)
- 1511.05644 (arxiv.org)
- 10.1109/ICCV.2017.304 (doi.org)
- 1611.04076 (arxiv.org)
- "Least Squares Generative Adversarial Networks" (thecvf.com)
- 1705.02894 (arxiv.org)
- "Generative Adversarial Networks (GANs), Presentation at Berkeley Artificial Intelligence Lab" (iangoodfellow.com)
- 1412.6515 (arxiv.org)
- 1701.04862 (arxiv.org)
- 1705.08868 (arxiv.org)
- 2102.07074 (arxiv.org)
- 1512.09300 (arxiv.org)
- "Autoencoding beyond pixels using a learned similarity metric" (mlr.press)
- "Self-Attention Generative Adversarial Networks" (mlr.press)
- 1610.09585 (arxiv.org)
- "Conditional Image Synthesis with Auxiliary Classifier GANs" (mlr.press)
- The GAN Zoo (github.com)
- 10.1016/j.cviu.2018.10.009 (doi.org)
- 1802.03446 (arxiv.org)
- 1801.03924 (arxiv.org)
- 1412.6980 (arxiv.org)
- 1706.08500 (arxiv.org)
- "GANs Trained by a Two Time-Scale Update Rule Converge to a Local Nash Equilibrium" (neurips.cc)
- 1809.11096 (arxiv.org)
- Large Scale GAN Training for High Fidelity Natural Image Synthesis (harvard.edu)
- 1801.04406 (arxiv.org)
- 1712.04086 (arxiv.org)
- "r/MachineLearning - Comment by u/ian_goodfellow on "[R] [1701.07875] Wasserstein GAN" (reddit.com)
- 1904.03626 (arxiv.org)
- "On The Power of Curriculum Learning in Training Deep Networks" (mlr.press)
- 1910.08967 (arxiv.org)
- cs.NE (arxiv.org)
- 1710.10196 (arxiv.org)
- 1904.08994 (arxiv.org)
- "Do GANs always have Nash equilibria?" (mlr.press)
- 1401.4082 (arxiv.org)
- "Stochastic Backpropagation and Approximate Inference in Deep Generative Models" (mlr.press)
- 1312.6114 (arxiv.org)
- 1701.00160 (arxiv.org)
- 1610.03483 (arxiv.org)
- Generative Models (openai.com)
- 1611.08408 (arxiv.org)
- "Vanilla GAN (GANs in computer vision: Introduction to generative learning)" (theaisummer.com)
- 1606.03476 (arxiv.org)
- "Generative Adversarial Imitation Learning" (nips.cc)
- "Image-to-Image Translation with Conditional Adversarial Nets" (phillipi.github.io)
- 1606.03498 (arxiv.org)
- Generative Adversarial Nets (nips.cc)
- "CSDL | IEEE Computer Society" (computer.org)
- 1906.01529 (arxiv.org)
- This Cat Does Not Exist (thiscatdoesnotexist.com)
- This Person Does Not Exist (thispersondoesnotexist.com)
- Generative Adversarial Network (GAN) (geeksforgeeks.org)
- Generative Adversarial Networks (medium.com)
- Train Generative Adversarial Network (GAN) (mathworks.com)
- Generative Adversarial Networks (GANs): Architecture and ... (leewayhertz.com)
Featured Images: pexels.com