medgan cgan b-gan ff lapgan mpm-gan generative · pdf fileian goodfellow, staff research...

24
Ian Goodfellow, StaResearch Scientist, Google Brain ICCV Tutorial on GANs Venice, 2017-10-22 Generative Adversarial Networks 3D-GAN AC-GAN AdaGAN AGAN AL-CGAN ALI AMGAN AnoGAN ArtGAN b-GAN Bayesian GAN BEGAN BiGAN BS-GAN CGAN CCGAN CatGAN CoGAN Context-RNN-GAN C-RNN-GAN C-VAE-GAN CycleGAN DTN DCGAN DiscoGAN DR-GAN DualGAN EBGAN f-GAN FF-GAN GAWWN GoGAN GP-GAN IAN iGAN IcGAN ID-CGAN InfoGAN LAPGAN LR-GAN LS-GAN LSGAN MGAN MAGAN MAD-GAN MalGAN MARTA-GAN McGAN MedGAN MIX+GAN MPM-GAN GMAN alpha-GAN WGAN-GP

Upload: trinhnga

Post on 01-Feb-2018

239 views

Category:

Documents


6 download

TRANSCRIPT

Page 1: MedGAN CGAN b-GAN ff LAPGAN MPM-GAN Generative · PDF fileIan Goodfellow, Staff Research Scientist, Google Brain ICCV Tutorial on GANs Venice, 2017-10-22 Generative Adversarial Networks

Ian Goodfellow, Staff Research Scientist, Google Brain ICCV Tutorial on GANs

Venice, 2017-10-22

Generative Adversarial Networks

3D-GAN AC-GAN

AdaGANAffGAN

AL-CGANALI

AMGAN

AnoGAN

ArtGAN

b-GAN

Bayesian GAN

BEGAN

BiGAN

BS-GAN

CGAN

CCGAN

CatGAN

CoGAN

Context-RNN-GAN

C-RNN-GANC-VAE-GAN

CycleGAN

DTN

DCGAN

DiscoGAN

DR-GAN

DualGAN

EBGAN

f-GAN

FF-GAN

GAWWN

GoGAN

GP-GAN

IANiGAN

IcGANID-CGAN

InfoGANLAPGAN

LR-GANLS-GAN

LSGAN

MGAN

MAGAN

MAD-GAN

MalGANMARTA-GAN

McGAN

MedGAN

MIX+GAN

MPM-GAN

GMANalpha-GAN

WGAN-GP

Page 2: MedGAN CGAN b-GAN ff LAPGAN MPM-GAN Generative · PDF fileIan Goodfellow, Staff Research Scientist, Google Brain ICCV Tutorial on GANs Venice, 2017-10-22 Generative Adversarial Networks

(Goodfellow 2017)

Generative Modeling• Density estimation

• Sample generation

Training examples Model samples

Page 3: MedGAN CGAN b-GAN ff LAPGAN MPM-GAN Generative · PDF fileIan Goodfellow, Staff Research Scientist, Google Brain ICCV Tutorial on GANs Venice, 2017-10-22 Generative Adversarial Networks

(Goodfellow 2017)

Maximum Likelihood

BRIEF ARTICLE

THE AUTHOR

⇤= argmax

✓Ex⇠pdata log pmodel

(x | ✓)

1

Page 4: MedGAN CGAN b-GAN ff LAPGAN MPM-GAN Generative · PDF fileIan Goodfellow, Staff Research Scientist, Google Brain ICCV Tutorial on GANs Venice, 2017-10-22 Generative Adversarial Networks

(Goodfellow 2017)

Adversarial Nets Framework

x sampled from data

Differentiable function D

D(x) tries to be near 1

Input noise z

Differentiable function G

x sampled from model

D

D tries to make D(G(z)) near 0,G tries to make D(G(z)) near 1

(Goodfellow et al., 2014)

Page 5: MedGAN CGAN b-GAN ff LAPGAN MPM-GAN Generative · PDF fileIan Goodfellow, Staff Research Scientist, Google Brain ICCV Tutorial on GANs Venice, 2017-10-22 Generative Adversarial Networks

(Goodfellow 2017)

What can you do with GANs?• Simulated environments and training data

• Missing data

• Semi-supervised learning

• Multiple correct answers

• Realistic generation tasks

• Simulation by prediction

• Solve inference problems

• Learn useful embeddings

Page 6: MedGAN CGAN b-GAN ff LAPGAN MPM-GAN Generative · PDF fileIan Goodfellow, Staff Research Scientist, Google Brain ICCV Tutorial on GANs Venice, 2017-10-22 Generative Adversarial Networks

(Goodfellow 2017)

Page 7: MedGAN CGAN b-GAN ff LAPGAN MPM-GAN Generative · PDF fileIan Goodfellow, Staff Research Scientist, Google Brain ICCV Tutorial on GANs Venice, 2017-10-22 Generative Adversarial Networks

(Goodfellow 2017)

GANs for simulated training data

(Shrivastava et al., 2016)

Page 8: MedGAN CGAN b-GAN ff LAPGAN MPM-GAN Generative · PDF fileIan Goodfellow, Staff Research Scientist, Google Brain ICCV Tutorial on GANs Venice, 2017-10-22 Generative Adversarial Networks

(Raffel, 2017)

GANs for domain adaptation

(Bousmalis et al., 2016)

Page 9: MedGAN CGAN b-GAN ff LAPGAN MPM-GAN Generative · PDF fileIan Goodfellow, Staff Research Scientist, Google Brain ICCV Tutorial on GANs Venice, 2017-10-22 Generative Adversarial Networks

(Goodfellow 2017)

What can you do with GANs?• Simulated environments and training data

• Missing data

• Semi-supervised learning

• Multiple correct answers

• Realistic generation tasks

• Simulation by prediction

• Solve inference problems

• Learn useful embeddings

Page 10: MedGAN CGAN b-GAN ff LAPGAN MPM-GAN Generative · PDF fileIan Goodfellow, Staff Research Scientist, Google Brain ICCV Tutorial on GANs Venice, 2017-10-22 Generative Adversarial Networks

(Goodfellow 2017)

Generative modeling reveals a face

(Yeh et al., 2016)

Page 11: MedGAN CGAN b-GAN ff LAPGAN MPM-GAN Generative · PDF fileIan Goodfellow, Staff Research Scientist, Google Brain ICCV Tutorial on GANs Venice, 2017-10-22 Generative Adversarial Networks

(Goodfellow 2017)

What can you do with GANs?• Simulated environments and training data

• Missing data

• Semi-supervised learning

• Multiple correct answers

• Realistic generation tasks

• Simulation by prediction

• Solve inference problems

• Learn useful embeddings

Page 12: MedGAN CGAN b-GAN ff LAPGAN MPM-GAN Generative · PDF fileIan Goodfellow, Staff Research Scientist, Google Brain ICCV Tutorial on GANs Venice, 2017-10-22 Generative Adversarial Networks

(Goodfellow 2017)

Supervised Discriminator

Input

Real

Hidden units

Fake

Input

Real dog

Hidden units

FakeReal cat

(Odena 2016, Salimans et al 2016)

Page 13: MedGAN CGAN b-GAN ff LAPGAN MPM-GAN Generative · PDF fileIan Goodfellow, Staff Research Scientist, Google Brain ICCV Tutorial on GANs Venice, 2017-10-22 Generative Adversarial Networks

(Goodfellow 2017)

What can you do with GANs?• Simulated environments and training data

• Missing data

• Semi-supervised learning

• Multiple correct answers

• Realistic generation tasks

• Simulation by prediction

• Solve inference problems

• Learn useful embeddings

Page 14: MedGAN CGAN b-GAN ff LAPGAN MPM-GAN Generative · PDF fileIan Goodfellow, Staff Research Scientist, Google Brain ICCV Tutorial on GANs Venice, 2017-10-22 Generative Adversarial Networks

(Goodfellow 2017)

Next Video Frame PredictionCHAPTER 15. REPRESENTATION LEARNING

Ground Truth MSE Adversarial

Figure 15.6: Predictive generative networks provide an example of the importance oflearning which features are salient. In this example, the predictive generative networkhas been trained to predict the appearance of a 3-D model of a human head at a specificviewing angle. (Left)Ground truth. This is the correct image, that the network shouldemit. (Center)Image produced by a predictive generative network trained with meansquared error alone. Because the ears do not cause an extreme difference in brightnesscompared to the neighboring skin, they were not sufficiently salient for the model to learnto represent them. (Right)Image produced by a model trained with a combination ofmean squared error and adversarial loss. Using this learned cost function, the ears aresalient because they follow a predictable pattern. Learning which underlying causes areimportant and relevant enough to model is an important active area of research. Figuresgraciously provided by Lotter et al. (2015).

recognizable shape and consistent position means that a feedforward networkcan easily learn to detect them, making them highly salient under the generativeadversarial framework. See figure 15.6 for example images. Generative adversarialnetworks are only one step toward determining which factors should be represented.We expect that future research will discover better ways of determining whichfactors to represent, and develop mechanisms for representing different factorsdepending on the task.

A benefit of learning the underlying causal factors, as pointed out by Schölkopfet al. (2012), is that if the true generative process has x as an effect and y asa cause, then modeling p(x | y) is robust to changes in p(y). If the cause-effectrelationship was reversed, this would not be true, since by Bayes’ rule, p(x | y)

would be sensitive to changes in p(y). Very often, when we consider changes indistribution due to different domains, temporal non-stationarity, or changes inthe nature of the task, the causal mechanisms remain invariant (the laws of theuniverse are constant) while the marginal distribution over the underlying causescan change. Hence, better generalization and robustness to all kinds of changes can

545

(Lotter et al 2016)

What happens next?

Page 15: MedGAN CGAN b-GAN ff LAPGAN MPM-GAN Generative · PDF fileIan Goodfellow, Staff Research Scientist, Google Brain ICCV Tutorial on GANs Venice, 2017-10-22 Generative Adversarial Networks

(Goodfellow 2017)

CHAPTER 15. REPRESENTATION LEARNING

Ground Truth MSE Adversarial

Figure 15.6: Predictive generative networks provide an example of the importance oflearning which features are salient. In this example, the predictive generative networkhas been trained to predict the appearance of a 3-D model of a human head at a specificviewing angle. (Left)Ground truth. This is the correct image, that the network shouldemit. (Center)Image produced by a predictive generative network trained with meansquared error alone. Because the ears do not cause an extreme difference in brightnesscompared to the neighboring skin, they were not sufficiently salient for the model to learnto represent them. (Right)Image produced by a model trained with a combination ofmean squared error and adversarial loss. Using this learned cost function, the ears aresalient because they follow a predictable pattern. Learning which underlying causes areimportant and relevant enough to model is an important active area of research. Figuresgraciously provided by Lotter et al. (2015).

recognizable shape and consistent position means that a feedforward networkcan easily learn to detect them, making them highly salient under the generativeadversarial framework. See figure 15.6 for example images. Generative adversarialnetworks are only one step toward determining which factors should be represented.We expect that future research will discover better ways of determining whichfactors to represent, and develop mechanisms for representing different factorsdepending on the task.

A benefit of learning the underlying causal factors, as pointed out by Schölkopfet al. (2012), is that if the true generative process has x as an effect and y asa cause, then modeling p(x | y) is robust to changes in p(y). If the cause-effectrelationship was reversed, this would not be true, since by Bayes’ rule, p(x | y)

would be sensitive to changes in p(y). Very often, when we consider changes indistribution due to different domains, temporal non-stationarity, or changes inthe nature of the task, the causal mechanisms remain invariant (the laws of theuniverse are constant) while the marginal distribution over the underlying causescan change. Hence, better generalization and robustness to all kinds of changes can

545

Next Video Frame Prediction

(Lotter et al 2016)

Page 16: MedGAN CGAN b-GAN ff LAPGAN MPM-GAN Generative · PDF fileIan Goodfellow, Staff Research Scientist, Google Brain ICCV Tutorial on GANs Venice, 2017-10-22 Generative Adversarial Networks

(Raffel, 2017)

Next Video Frame(s) Prediction

(Mathieu et al. 2015)

Mean Squared Error Mean Absolute Error Adversarial

Page 17: MedGAN CGAN b-GAN ff LAPGAN MPM-GAN Generative · PDF fileIan Goodfellow, Staff Research Scientist, Google Brain ICCV Tutorial on GANs Venice, 2017-10-22 Generative Adversarial Networks

(Goodfellow 2017)

What can you do with GANs?• Simulated environments and training data

• Missing data

• Semi-supervised learning

• Multiple correct answers

• Realistic generation tasks

• Simulation by prediction

• Solve inference problems

• Learn useful embeddings

Page 18: MedGAN CGAN b-GAN ff LAPGAN MPM-GAN Generative · PDF fileIan Goodfellow, Staff Research Scientist, Google Brain ICCV Tutorial on GANs Venice, 2017-10-22 Generative Adversarial Networks

(Goodfellow 2017)

Which of these are real photos ?

(work by vue.ai covered by Quartz)

Page 19: MedGAN CGAN b-GAN ff LAPGAN MPM-GAN Generative · PDF fileIan Goodfellow, Staff Research Scientist, Google Brain ICCV Tutorial on GANs Venice, 2017-10-22 Generative Adversarial Networks

(Goodfellow 2017)

What can you do with GANs?• Simulated environments and training data

• Missing data

• Semi-supervised learning

• Multiple correct answers

• Realistic generation tasks

• Simulation by prediction

• Solve inference problems

• Learn useful embeddings

Page 20: MedGAN CGAN b-GAN ff LAPGAN MPM-GAN Generative · PDF fileIan Goodfellow, Staff Research Scientist, Google Brain ICCV Tutorial on GANs Venice, 2017-10-22 Generative Adversarial Networks

(Goodfellow 2017)

Vector Space ArithmeticCHAPTER 15. REPRESENTATION LEARNING

- + =

Figure 15.9: A generative model has learned a distributed representation that disentanglesthe concept of gender from the concept of wearing glasses. If we begin with the repre-sentation of the concept of a man with glasses, then subtract the vector representing theconcept of a man without glasses, and finally add the vector representing the conceptof a woman without glasses, we obtain the vector representing the concept of a womanwith glasses. The generative model correctly decodes all of these representation vectors toimages that may be recognized as belonging to the correct class. Images reproduced withpermission from Radford et al. (2015).

common is that one could imagine learning about each of them without having to

see all the configurations of all the others. Radford et al. (2015) demonstrated thata generative model can learn a representation of images of faces, with separatedirections in representation space capturing different underlying factors of variation.Figure 15.9 demonstrates that one direction in representation space correspondsto whether the person is male or female, while another corresponds to whetherthe person is wearing glasses. These features were discovered automatically, notfixed a priori. There is no need to have labels for the hidden unit classifiers:gradient descent on an objective function of interest naturally learns semanticallyinteresting features, so long as the task requires such features. We can learn aboutthe distinction between male and female, or about the presence or absence ofglasses, without having to characterize all of the configurations of the n � 1 otherfeatures by examples covering all of these combinations of values. This form ofstatistical separability is what allows one to generalize to new configurations of aperson’s features that have never been seen during training.

552

Man with glasses

Man Woman

Woman with Glasses

(Radford et al, 2015)

Page 21: MedGAN CGAN b-GAN ff LAPGAN MPM-GAN Generative · PDF fileIan Goodfellow, Staff Research Scientist, Google Brain ICCV Tutorial on GANs Venice, 2017-10-22 Generative Adversarial Networks

(Goodfellow 2017)

How long until GANs can do this?

Training examples Model samples

Page 22: MedGAN CGAN b-GAN ff LAPGAN MPM-GAN Generative · PDF fileIan Goodfellow, Staff Research Scientist, Google Brain ICCV Tutorial on GANs Venice, 2017-10-22 Generative Adversarial Networks

(Goodfellow 2017)

AC-GANs

(Odena et al., 2016)

Page 23: MedGAN CGAN b-GAN ff LAPGAN MPM-GAN Generative · PDF fileIan Goodfellow, Staff Research Scientist, Google Brain ICCV Tutorial on GANs Venice, 2017-10-22 Generative Adversarial Networks

(Goodfellow 2017)

Track updates at the GAN Zoo

https://github.com/hindupuravinash/the-gan-zoo

Page 24: MedGAN CGAN b-GAN ff LAPGAN MPM-GAN Generative · PDF fileIan Goodfellow, Staff Research Scientist, Google Brain ICCV Tutorial on GANs Venice, 2017-10-22 Generative Adversarial Networks

(Goodfellow 2017)

Questions?