SlideShare a Scribd company logo
Kevin McGuinness
kevin.mcguinness@dcu.ie
Assistant Professor
School of Electronic Engineering
Dublin City University
http://bit.ly/dlcv2018
#DLUPC
Generative Models
Day 4 Lecture 3
What is a generative model?
A model P(X; ϴ) that we can draw samples
from.
E.g. A Gaussian Mixture Model
● Fitting: EM algorithm
● Drawing samples:
○ Draw sample from categorical distribution to
select Gaussian
○ Draw sample from Gaussian
GMMs are not generally complex enough
to draw samples of images from.
P(X = x)
x
x
2
Why are generative models important?
● Model the probability density of images
● Understanding P(X) may help us understand P(Y | X)
● Generate novel content
● Generate training data for discriminative networks
● Artistic applications
● Image completion
● Monte-carlo estimators
3
Generative adversarial networks
Novel method of training deep generative models invented by Ian Goodfellow et al.
in 2014
Idea: pit a generator and a discriminator against each other
● Generator tries to draw samples from P(X)
● Discriminator tries to tell if sample came from the generator or the real world
Both discriminator and generator are deep networks (differentiable functions)
Can train with backprop: train discriminator for a while, then train generator, then
discriminator, …
4
Generative adversarial networks (conceptual)
Generator
Real world
images
Discriminator
Real
Loss
Latentrandomvariable
Sample
Sample
Fake
5
The generator
Deterministic mapping from a latent random vector to sample from q(x) ~ p(x)
Usually a deep neural network.
E.g. DCGAN:
6
The discriminator
Parameterised function that tries to distinguish between samples from real images
p(x) and generated ones q(x).
Usually a deep convolutional neural network.
conv
conv
...
F F
7
Training GANs
Generator
Real world
images
Discriminator
Real
Loss
Latentrandomvariable
Sample
Sample
Fake
Alternate between training the discriminator and generator
Differentiable module
Differentiable module
8
Generator
Real world
images
Discriminator
Real
Loss
Latentrandomvariable
Sample
Sample
Fake
1. Fix generator weights, draw samples from both real world and generated images
2. Train discriminator to distinguish between real world and generated images
Backprop error to
update discriminator
weights
9
Generator
Real world
images
Discriminator
Real
Loss
Latentrandomvariable
Sample
Sample
Fake
1. Fix discriminator weights
2. Sample from generator
3. Backprop error through discriminator to update generator weights
Backprop error to
update generator
weights
10
Training GANs
Iterate these two steps until convergence (which may not happen)
● Updating the discriminator should make it better at discriminating between real images and
generated ones (discriminator improves)
● Updating the generator makes it better at fooling the current discriminator (generator improves)
Eventually (we hope) that the generator gets so good that it is impossible for the discriminator to tell the
difference between real and generated images. Discriminator accuracy = 0.5
11
Discriminator
training
Generator
training
12
Some examples of generated images…
13
ImageNet
Source:
https://openai.com/blog/generative-models/
14
CIFAR-10
Source:
https://openai.com/blog/generative-models/
15
Credit:
Alec Radford
Code on GitHub 16
Credit: Alec Radford Code on GitHub 17
Issues
Known to be very difficult to train:
● Formulated as a “game” between two networks
● Unstable dynamics: hard to keep generator and discriminator in balance
● Optimization can oscillate between solutions
● Mode collapse in the generator
Difficult to evaluate results
18
Important variants
Wasserstein GAN (WGAN)
● MLE leads to a KL divergence loss.
● Numerical stability issues when estimated distribution and true distribution do not
overlap significantly (loss blows up).
● WGAN idea is to use a coarse approximation of the Wasserstein distance (the Earth
mover's distance).
● Weight clipping is needed to enforce Lipschitz constraint.
Overall effect is to make the GAN more stable. Discriminator can be trained more on
each step without blowing up.
Can work well in practice, but clipping the weights to enforce Lipschitz slows training.
Important variants
Least squares GAN (LSGAN)
● Similar motivation to WGAN: want a loss that gives nice gradients and doesn't
blow up.
● LSGAN Idea: just use squared error (L2
distance)!
● Turns out this is the same as minimizing the Pearson 2
divergence.
Deep Generative Models - Kevin McGuinness - UPC Barcelona 2018
Important variants
Energy-based GAN (EBGAN)
● Instead of using a binary classifier as the discriminator D use an energy-based model (an
autoencoder)
● D models the image manifold since it is trained on real images
● Optimize to generate samples that have low energy
● Generator gets more signal from D
Important variants
Boundary Equilibrium GAN (BEGAN)
● Combines ideas from WGAN and
EBGAN
● BEGAN idea: matching the distributions
of the reconstruction losses can be a
suitable proxy for matching the data
distributions.
● Use Wasserstein distance
approximation to do this
● Includes mechanism for automatically
maintaining equilibrium
Conditional GANs
GANs can be conditioned on other info: e.g. a label
● z might capture random characteristics of the
data, variabilities of possible futures,
● c would condition the deterministic parts (label)
For details on ways to condition GANs: Ways of Conditioning Generative Adversarial Networks (Wack et al.) 24
Generating images/frames conditioned on captions
(Reed et al. 2016b) (Zhang et al. 2016)
25
Predicting the future with adversarial training
Want to train a model to predict the pixels in frame (t+K) from pixels in frame t.
Many possible futures for same frame
Using supervised loss like MSE results in blurry solutions: loss if minimized if
predictor averages over possibilities when predicting.
We really want a sample, not the mean
Adversarial training can solve this: easy for an adversary to detect blurry frames
Mathieu et al. Deep multi-scale video prediction beyond mean square error, ICLR 2016 (https://arxiv.org/abs/1511.05440) 26
Mathieu et al. Deep multi-scale video prediction beyond mean square error, ICLR 2016 (https://arxiv.org/abs/1511.05440) 27
Image super-resolution
Bicubic: not using data statistics. SRResNet: trained with MSE. SRGAN is able to
understand that there are multiple correct answers, rather than averaging.
(Ledig et al. 2016)
28
Saliency prediction
Adversarial lossDala loss
Junting Pan, Cristian Canton, Kevin McGuinness, Noel E. O’Connor, Jordi Torres, Elisa Sayrol and Xavier
Giro-i-Nieto. “SalGAN: Visual Saliency Prediction with Generative Adversarial Networks.” arXiv. 2017.
29
Image-to-Image translation
Isola, Phillip, Jun-Yan Zhu, Tinghui Zhou, and Alexei A. Efros. "Image-to-image translation with conditional adversarial networks."
arXiv:1611.07004 (2016).
Generator
Discriminator
Generated
pairs
Real World
Ground truth
pairs
Loss
30
Questions?

More Related Content

PDF
Semantic Segmentation - Míriam Bellver - UPC Barcelona 2018
PDF
Generative Models and Adversarial Training (D3L4 2017 UPC Deep Learning for ...
PDF
Deep 3D Analysis - Javier Ruiz-Hidalgo - UPC Barcelona 2018
PDF
Generative Models and Adversarial Training (D2L3 Insight@DCU Machine Learning...
PDF
Joint unsupervised learning of deep representations and image clusters
PDF
Deep Learning for Computer Vision: Attention Models (UPC 2016)
PDF
Unsupervised Deep Learning (D2L1 Insight@DCU Machine Learning Workshop 2017)
PDF
Interpretability of Convolutional Neural Networks - Eva Mohedano - UPC Barcel...
Semantic Segmentation - Míriam Bellver - UPC Barcelona 2018
Generative Models and Adversarial Training (D3L4 2017 UPC Deep Learning for ...
Deep 3D Analysis - Javier Ruiz-Hidalgo - UPC Barcelona 2018
Generative Models and Adversarial Training (D2L3 Insight@DCU Machine Learning...
Joint unsupervised learning of deep representations and image clusters
Deep Learning for Computer Vision: Attention Models (UPC 2016)
Unsupervised Deep Learning (D2L1 Insight@DCU Machine Learning Workshop 2017)
Interpretability of Convolutional Neural Networks - Eva Mohedano - UPC Barcel...

What's hot (20)

PDF
D1L5 Visualization (D1L2 Insight@DCU Machine Learning Workshop 2017)
PDF
Object Segmentation (D2L7 Insight@DCU Machine Learning Workshop 2017)
PDF
Deep 3D Visual Analysis - Javier Ruiz-Hidalgo - UPC Barcelona 2017
PDF
Generative Adversarial Networks GAN - Santiago Pascual - UPC Barcelona 2018
PDF
Optimizing Deep Networks (D1L6 Insight@DCU Machine Learning Workshop 2017)
PDF
Image Segmentation (D3L1 2017 UPC Deep Learning for Computer Vision)
PDF
Deep Learning for Computer Vision: Segmentation (UPC 2016)
PDF
Attention Models (D3L6 2017 UPC Deep Learning for Computer Vision)
PDF
Training Deep Networks with Backprop (D1L4 Insight@DCU Machine Learning Works...
PDF
Deep Learning for Computer Vision: Generative models and adversarial training...
PDF
Deep Learning for Computer Vision: Visualization (UPC 2016)
PDF
Unsupervised Learning (D2L6 2017 UPC Deep Learning for Computer Vision)
PDF
Deep Learning for Computer Vision: Unsupervised Learning (UPC 2016)
PDF
Variational Autoencoders VAE - Santiago Pascual - UPC Barcelona 2018
PDF
Recurrent Instance Segmentation (UPC Reading Group)
PDF
Visualization of Deep Learning Models (D1L6 2017 UPC Deep Learning for Comput...
PDF
Convolutional Neural Networks (D1L3 2017 UPC Deep Learning for Computer Vision)
PDF
Perceptrons (D1L2 2017 UPC Deep Learning for Computer Vision)
PDF
Intepretability / Explainable AI for Deep Neural Networks
PDF
crfasrnn_presentation
D1L5 Visualization (D1L2 Insight@DCU Machine Learning Workshop 2017)
Object Segmentation (D2L7 Insight@DCU Machine Learning Workshop 2017)
Deep 3D Visual Analysis - Javier Ruiz-Hidalgo - UPC Barcelona 2017
Generative Adversarial Networks GAN - Santiago Pascual - UPC Barcelona 2018
Optimizing Deep Networks (D1L6 Insight@DCU Machine Learning Workshop 2017)
Image Segmentation (D3L1 2017 UPC Deep Learning for Computer Vision)
Deep Learning for Computer Vision: Segmentation (UPC 2016)
Attention Models (D3L6 2017 UPC Deep Learning for Computer Vision)
Training Deep Networks with Backprop (D1L4 Insight@DCU Machine Learning Works...
Deep Learning for Computer Vision: Generative models and adversarial training...
Deep Learning for Computer Vision: Visualization (UPC 2016)
Unsupervised Learning (D2L6 2017 UPC Deep Learning for Computer Vision)
Deep Learning for Computer Vision: Unsupervised Learning (UPC 2016)
Variational Autoencoders VAE - Santiago Pascual - UPC Barcelona 2018
Recurrent Instance Segmentation (UPC Reading Group)
Visualization of Deep Learning Models (D1L6 2017 UPC Deep Learning for Comput...
Convolutional Neural Networks (D1L3 2017 UPC Deep Learning for Computer Vision)
Perceptrons (D1L2 2017 UPC Deep Learning for Computer Vision)
Intepretability / Explainable AI for Deep Neural Networks
crfasrnn_presentation
Ad

Similar to Deep Generative Models - Kevin McGuinness - UPC Barcelona 2018 (20)

PDF
Deep Generative Models II (DLAI D10L1 2017 UPC Deep Learning for Artificial I...
PDF
Tutorial on Theory and Application of Generative Adversarial Networks
PDF
A Short Introduction to Generative Adversarial Networks
PDF
gans_copy.pdfhjsjsisidkskskkskwkduydjekedj
PDF
Introduction to Generative Adversarial Network
PDF
Generative Adversarial Networks GAN - Xavier Giro - UPC TelecomBCN Barcelona ...
PPTX
ICASSP 2018 Tutorial: Generative Adversarial Network and its Applications to ...
PDF
PPTX
GAN Deep Learning Approaches to Image Processing Applications (1).pptx
PPTX
Module4_GAN.pptxgdgdijehejejjejejejhehjdd
PPTX
GAN_SANTHOSH KUMAR_Architecture_in_network
PDF
Generative adversarial networks
PDF
1시간만에 GAN(Generative Adversarial Network) 완전 정복하기
PPTX
Anime_face_generation_through_DCGAN.pptx
PDF
Deep Generative Modelling (updated)
PDF
Generative adversarial network_Ayadi_Alaeddine
PDF
PDF
11_gan.pdf
PDF
Generative Adversarial Networks (D2L5 Deep Learning for Speech and Language U...
PDF
Jakub Langr (University of Oxford) - Overview of Generative Adversarial Netwo...
Deep Generative Models II (DLAI D10L1 2017 UPC Deep Learning for Artificial I...
Tutorial on Theory and Application of Generative Adversarial Networks
A Short Introduction to Generative Adversarial Networks
gans_copy.pdfhjsjsisidkskskkskwkduydjekedj
Introduction to Generative Adversarial Network
Generative Adversarial Networks GAN - Xavier Giro - UPC TelecomBCN Barcelona ...
ICASSP 2018 Tutorial: Generative Adversarial Network and its Applications to ...
GAN Deep Learning Approaches to Image Processing Applications (1).pptx
Module4_GAN.pptxgdgdijehejejjejejejhehjdd
GAN_SANTHOSH KUMAR_Architecture_in_network
Generative adversarial networks
1시간만에 GAN(Generative Adversarial Network) 완전 정복하기
Anime_face_generation_through_DCGAN.pptx
Deep Generative Modelling (updated)
Generative adversarial network_Ayadi_Alaeddine
11_gan.pdf
Generative Adversarial Networks (D2L5 Deep Learning for Speech and Language U...
Jakub Langr (University of Oxford) - Overview of Generative Adversarial Netwo...
Ad

More from Universitat Politècnica de Catalunya (20)

PDF
Deep Generative Learning for All - The Gen AI Hype (Spring 2024)
PDF
Deep Generative Learning for All
PDF
The Transformer in Vision | Xavier Giro | Master in Computer Vision Barcelona...
PDF
Towards Sign Language Translation & Production | Xavier Giro-i-Nieto
PDF
The Transformer - Xavier Giró - UPC Barcelona 2021
PDF
Learning Representations for Sign Language Videos - Xavier Giro - NIST TRECVI...
PDF
Open challenges in sign language translation and production
PPTX
Generation of Synthetic Referring Expressions for Object Segmentation in Videos
PPTX
Discovery and Learning of Navigation Goals from Pixels in Minecraft
PDF
Learn2Sign : Sign language recognition and translation using human keypoint e...
PDF
Convolutional Neural Networks - Xavier Giro - UPC TelecomBCN Barcelona 2020
PDF
Self-Supervised Audio-Visual Learning - Xavier Giro - UPC TelecomBCN Barcelon...
PDF
Attention for Deep Learning - Xavier Giro - UPC TelecomBCN Barcelona 2020
PDF
Q-Learning with a Neural Network - Xavier Giró - UPC Barcelona 2020
PDF
Language and Vision with Deep Learning - Xavier Giró - ACM ICMR 2020 (Tutorial)
PDF
Image Segmentation with Deep Learning - Xavier Giro & Carles Ventura - ISSonD...
PDF
Curriculum Learning for Recurrent Video Object Segmentation
PDF
Deep Self-supervised Learning for All - Xavier Giro - X-Europe 2020
PDF
Deep Learning Representations for All - Xavier Giro-i-Nieto - IRI Barcelona 2020
PDF
Transcription-Enriched Joint Embeddings for Spoken Descriptions of Images and...
Deep Generative Learning for All - The Gen AI Hype (Spring 2024)
Deep Generative Learning for All
The Transformer in Vision | Xavier Giro | Master in Computer Vision Barcelona...
Towards Sign Language Translation & Production | Xavier Giro-i-Nieto
The Transformer - Xavier Giró - UPC Barcelona 2021
Learning Representations for Sign Language Videos - Xavier Giro - NIST TRECVI...
Open challenges in sign language translation and production
Generation of Synthetic Referring Expressions for Object Segmentation in Videos
Discovery and Learning of Navigation Goals from Pixels in Minecraft
Learn2Sign : Sign language recognition and translation using human keypoint e...
Convolutional Neural Networks - Xavier Giro - UPC TelecomBCN Barcelona 2020
Self-Supervised Audio-Visual Learning - Xavier Giro - UPC TelecomBCN Barcelon...
Attention for Deep Learning - Xavier Giro - UPC TelecomBCN Barcelona 2020
Q-Learning with a Neural Network - Xavier Giró - UPC Barcelona 2020
Language and Vision with Deep Learning - Xavier Giró - ACM ICMR 2020 (Tutorial)
Image Segmentation with Deep Learning - Xavier Giro & Carles Ventura - ISSonD...
Curriculum Learning for Recurrent Video Object Segmentation
Deep Self-supervised Learning for All - Xavier Giro - X-Europe 2020
Deep Learning Representations for All - Xavier Giro-i-Nieto - IRI Barcelona 2020
Transcription-Enriched Joint Embeddings for Spoken Descriptions of Images and...

Recently uploaded (20)

PPTX
Business Ppt On Nestle.pptx huunnnhhgfvu
PPTX
Acceptance and paychological effects of mandatory extra coach I classes.pptx
PPT
Quality review (1)_presentation of this 21
PPTX
1_Introduction to advance data techniques.pptx
PDF
Galatica Smart Energy Infrastructure Startup Pitch Deck
PPTX
Introduction to Knowledge Engineering Part 1
PDF
Foundation of Data Science unit number two notes
PDF
.pdf is not working space design for the following data for the following dat...
PDF
Launch Your Data Science Career in Kochi – 2025
PPT
Reliability_Chapter_ presentation 1221.5784
PPTX
DISORDERS OF THE LIVER, GALLBLADDER AND PANCREASE (1).pptx
PPTX
MODULE 8 - DISASTER risk PREPAREDNESS.pptx
PPTX
Data_Analytics_and_PowerBI_Presentation.pptx
PDF
Lecture1 pattern recognition............
PPTX
mbdjdhjjodule 5-1 rhfhhfjtjjhafbrhfnfbbfnb
PPT
Chapter 3 METAL JOINING.pptnnnnnnnnnnnnn
PPTX
Introduction to Basics of Ethical Hacking and Penetration Testing -Unit No. 1...
PPTX
Global journeys: estimating international migration
PPTX
iec ppt-1 pptx icmr ppt on rehabilitation.pptx
Business Ppt On Nestle.pptx huunnnhhgfvu
Acceptance and paychological effects of mandatory extra coach I classes.pptx
Quality review (1)_presentation of this 21
1_Introduction to advance data techniques.pptx
Galatica Smart Energy Infrastructure Startup Pitch Deck
Introduction to Knowledge Engineering Part 1
Foundation of Data Science unit number two notes
.pdf is not working space design for the following data for the following dat...
Launch Your Data Science Career in Kochi – 2025
Reliability_Chapter_ presentation 1221.5784
DISORDERS OF THE LIVER, GALLBLADDER AND PANCREASE (1).pptx
MODULE 8 - DISASTER risk PREPAREDNESS.pptx
Data_Analytics_and_PowerBI_Presentation.pptx
Lecture1 pattern recognition............
mbdjdhjjodule 5-1 rhfhhfjtjjhafbrhfnfbbfnb
Chapter 3 METAL JOINING.pptnnnnnnnnnnnnn
Introduction to Basics of Ethical Hacking and Penetration Testing -Unit No. 1...
Global journeys: estimating international migration
iec ppt-1 pptx icmr ppt on rehabilitation.pptx

Deep Generative Models - Kevin McGuinness - UPC Barcelona 2018

  • 1. Kevin McGuinness kevin.mcguinness@dcu.ie Assistant Professor School of Electronic Engineering Dublin City University http://bit.ly/dlcv2018 #DLUPC Generative Models Day 4 Lecture 3
  • 2. What is a generative model? A model P(X; ϴ) that we can draw samples from. E.g. A Gaussian Mixture Model ● Fitting: EM algorithm ● Drawing samples: ○ Draw sample from categorical distribution to select Gaussian ○ Draw sample from Gaussian GMMs are not generally complex enough to draw samples of images from. P(X = x) x x 2
  • 3. Why are generative models important? ● Model the probability density of images ● Understanding P(X) may help us understand P(Y | X) ● Generate novel content ● Generate training data for discriminative networks ● Artistic applications ● Image completion ● Monte-carlo estimators 3
  • 4. Generative adversarial networks Novel method of training deep generative models invented by Ian Goodfellow et al. in 2014 Idea: pit a generator and a discriminator against each other ● Generator tries to draw samples from P(X) ● Discriminator tries to tell if sample came from the generator or the real world Both discriminator and generator are deep networks (differentiable functions) Can train with backprop: train discriminator for a while, then train generator, then discriminator, … 4
  • 5. Generative adversarial networks (conceptual) Generator Real world images Discriminator Real Loss Latentrandomvariable Sample Sample Fake 5
  • 6. The generator Deterministic mapping from a latent random vector to sample from q(x) ~ p(x) Usually a deep neural network. E.g. DCGAN: 6
  • 7. The discriminator Parameterised function that tries to distinguish between samples from real images p(x) and generated ones q(x). Usually a deep convolutional neural network. conv conv ... F F 7
  • 8. Training GANs Generator Real world images Discriminator Real Loss Latentrandomvariable Sample Sample Fake Alternate between training the discriminator and generator Differentiable module Differentiable module 8
  • 9. Generator Real world images Discriminator Real Loss Latentrandomvariable Sample Sample Fake 1. Fix generator weights, draw samples from both real world and generated images 2. Train discriminator to distinguish between real world and generated images Backprop error to update discriminator weights 9
  • 10. Generator Real world images Discriminator Real Loss Latentrandomvariable Sample Sample Fake 1. Fix discriminator weights 2. Sample from generator 3. Backprop error through discriminator to update generator weights Backprop error to update generator weights 10
  • 11. Training GANs Iterate these two steps until convergence (which may not happen) ● Updating the discriminator should make it better at discriminating between real images and generated ones (discriminator improves) ● Updating the generator makes it better at fooling the current discriminator (generator improves) Eventually (we hope) that the generator gets so good that it is impossible for the discriminator to tell the difference between real and generated images. Discriminator accuracy = 0.5 11
  • 13. Some examples of generated images… 13
  • 17. Credit: Alec Radford Code on GitHub 17
  • 18. Issues Known to be very difficult to train: ● Formulated as a “game” between two networks ● Unstable dynamics: hard to keep generator and discriminator in balance ● Optimization can oscillate between solutions ● Mode collapse in the generator Difficult to evaluate results 18
  • 19. Important variants Wasserstein GAN (WGAN) ● MLE leads to a KL divergence loss. ● Numerical stability issues when estimated distribution and true distribution do not overlap significantly (loss blows up). ● WGAN idea is to use a coarse approximation of the Wasserstein distance (the Earth mover's distance). ● Weight clipping is needed to enforce Lipschitz constraint. Overall effect is to make the GAN more stable. Discriminator can be trained more on each step without blowing up. Can work well in practice, but clipping the weights to enforce Lipschitz slows training.
  • 20. Important variants Least squares GAN (LSGAN) ● Similar motivation to WGAN: want a loss that gives nice gradients and doesn't blow up. ● LSGAN Idea: just use squared error (L2 distance)! ● Turns out this is the same as minimizing the Pearson 2 divergence.
  • 22. Important variants Energy-based GAN (EBGAN) ● Instead of using a binary classifier as the discriminator D use an energy-based model (an autoencoder) ● D models the image manifold since it is trained on real images ● Optimize to generate samples that have low energy ● Generator gets more signal from D
  • 23. Important variants Boundary Equilibrium GAN (BEGAN) ● Combines ideas from WGAN and EBGAN ● BEGAN idea: matching the distributions of the reconstruction losses can be a suitable proxy for matching the data distributions. ● Use Wasserstein distance approximation to do this ● Includes mechanism for automatically maintaining equilibrium
  • 24. Conditional GANs GANs can be conditioned on other info: e.g. a label ● z might capture random characteristics of the data, variabilities of possible futures, ● c would condition the deterministic parts (label) For details on ways to condition GANs: Ways of Conditioning Generative Adversarial Networks (Wack et al.) 24
  • 25. Generating images/frames conditioned on captions (Reed et al. 2016b) (Zhang et al. 2016) 25
  • 26. Predicting the future with adversarial training Want to train a model to predict the pixels in frame (t+K) from pixels in frame t. Many possible futures for same frame Using supervised loss like MSE results in blurry solutions: loss if minimized if predictor averages over possibilities when predicting. We really want a sample, not the mean Adversarial training can solve this: easy for an adversary to detect blurry frames Mathieu et al. Deep multi-scale video prediction beyond mean square error, ICLR 2016 (https://arxiv.org/abs/1511.05440) 26
  • 27. Mathieu et al. Deep multi-scale video prediction beyond mean square error, ICLR 2016 (https://arxiv.org/abs/1511.05440) 27
  • 28. Image super-resolution Bicubic: not using data statistics. SRResNet: trained with MSE. SRGAN is able to understand that there are multiple correct answers, rather than averaging. (Ledig et al. 2016) 28
  • 29. Saliency prediction Adversarial lossDala loss Junting Pan, Cristian Canton, Kevin McGuinness, Noel E. O’Connor, Jordi Torres, Elisa Sayrol and Xavier Giro-i-Nieto. “SalGAN: Visual Saliency Prediction with Generative Adversarial Networks.” arXiv. 2017. 29
  • 30. Image-to-Image translation Isola, Phillip, Jun-Yan Zhu, Tinghui Zhou, and Alexei A. Efros. "Image-to-image translation with conditional adversarial networks." arXiv:1611.07004 (2016). Generator Discriminator Generated pairs Real World Ground truth pairs Loss 30