Lecture 12: Generative Models

Lecture 12: Generative Models

Lecture 12: Generative Models Fei-Fei Li, Ranjay Krishna, Danfei Xu Lecture 12 - 1 May 11, 2021 Administrative ● A3 is out. Due May 25. ● Milestone was due May 10th ○ Read website page for milestone requirements. ○ Need to Finish data preprocessing and initial results by then. ● Midterm and A2 grades will be out this week Fei-Fei Li, Ranjay Krishna, Danfei Xu Lecture 12 -2 May 11, 2021 Supervised vs Unsupervised Learning Supervised Learning Data: (x, y) x is data, y is label Goal: Learn a function to map x -> y Examples: Classification, regression, object detection, semantic segmentation, image captioning, etc. Fei-Fei Li, Ranjay Krishna, Danfei Xu Lecture 12 - 3 May 11, 2021 Supervised vs Unsupervised Learning Supervised Learning Data: (x, y) x is data, y is label Cat Goal: Learn a function to map x -> y Examples: Classification, regression, object detection, Classification semantic segmentation, image captioning, etc. This image is CC0 public domain Fei-Fei Li, Ranjay Krishna, Danfei Xu Lecture 12 - 4 May 11, 2021 Supervised vs Unsupervised Learning Supervised Learning Data: (x, y) x is data, y is label Goal: Learn a function to map x -> y Examples: Classification, A cat sitting on a suitcase on the floor regression, object detection, semantic segmentation, image Image captioning captioning, etc. Caption generated using neuraltalk2 Image is CC0 Public domain. Fei-Fei Li, Ranjay Krishna, Danfei Xu Lecture 12 - 5 May 11, 2021 Supervised vs Unsupervised Learning Supervised Learning Data: (x, y) x is data, y is label Goal: Learn a function to map x -> y Examples: Classification, DOG, DOG, CAT regression, object detection, semantic segmentation, image Object Detection captioning, etc. This image is CC0 public domain Fei-Fei Li, Ranjay Krishna, Danfei Xu Lecture 12 - 6 May 11, 2021 Supervised vs Unsupervised Learning Supervised Learning Data: (x, y) x is data, y is label Goal: Learn a function to map x -> y Examples: Classification, GRASS, CAT, TREE, SKY regression, object detection, semantic segmentation, image Semantic Segmentation captioning, etc. Fei-Fei Li, Ranjay Krishna, Danfei Xu Lecture 12 - 7 May 11, 2021 Supervised vs Unsupervised Learning Unsupervised Learning Data: x Just data, no labels! Goal: Learn some underlying hidden structure of the data Examples: Clustering, dimensionality reduction, feature learning, density estimation, etc. Fei-Fei Li, Ranjay Krishna, Danfei Xu Lecture 12 - 8 May 11, 2021 Supervised vs Unsupervised Learning Unsupervised Learning Data: x Just data, no labels! Goal: Learn some underlying hidden structure of the data Examples: Clustering, K-means clustering dimensionality reduction, density estimation, etc. This image is CC0 public domain Fei-Fei Li, Ranjay Krishna, Danfei Xu Lecture 12 - 9 May 11, 2021 Supervised vs Unsupervised Learning Unsupervised Learning Data: x Just data, no labels! Goal: Learn some underlying hidden structure of the data 3-d 2-d Examples: Clustering, Principal Component Analysis dimensionality reduction, density (Dimensionality reduction) estimation, etc. This image from Matthias Scholz is CC0 public domain Fei-Fei Li, Ranjay Krishna, Danfei Xu Lecture 12 - 10 May 11, 2021 Supervised vs Unsupervised Learning Unsupervised Learning Data: x Figure copyright Ian Goodfellow, 2016. Reproduced with permission. Just data, no labels! 1-d density estimation Goal: Learn some underlying hidden structure of the data Examples: Clustering, dimensionality reduction, density 2-d density estimation estimation, etc. Modeling p(x) 2-d density images left and right are CC0 public domain Fei-Fei Li, Ranjay Krishna, Danfei Xu Lecture 12 - 11 May 11, 2021 Supervised vs Unsupervised Learning Supervised Learning Unsupervised Learning Data: (x, y) Data: x x is data, y is label Just data, no labels! Goal: Learn a function to map x -> y Goal: Learn some underlying hidden structure of the data Examples: Classification, regression, object detection, Examples: Clustering, semantic segmentation, image dimensionality reduction, density captioning, etc. estimation, etc. Fei-Fei Li, Ranjay Krishna, Danfei Xu Lecture 12 - 12 May 11, 2021 Generative Modeling Given training data, generate new samples from same distribution learning sampling pmodel(x ) Training data ~ pdata(x) Objectives: 1. Learn pmodel(x) that approximates pdata(x) 2. Sampling new x from pmodel(x) Fei-Fei Li, Ranjay Krishna, Danfei Xu Lecture 12 - 13 May 11, 2021 Generative Modeling Given training data, generate new samples from same distribution learning sampling pmodel(x ) Training data ~ pdata(x) Formulate as density estimation problems: - Explicit density estimation: explicitly define and solve for pmodel(x) - Implicit density estimation: learn model that can sample from pmodel(x) without explicitly defining it. Fei-Fei Li, Ranjay Krishna, Danfei Xu Lecture 12 - 14 May 11, 2021 Why Generative Models? - Realistic samples for artwork, super-resolution, colorization, etc. - Learn useful features for downstream tasks such as classification. - Getting insights from high-dimensional data (physics, medical imaging, etc.) - Modeling physical world for simulation and planning (robotics and reinforcement learning applications) - Many more ... FIgures from L-R are copyright: (1) Alec Radford et al. 2016; (2) Phillip Isola et al. 2017. Reproduced with authors permission (3) BAIR Blog. Fei-Fei Li, Ranjay Krishna, Danfei Xu Lecture 12 - 15 May 11, 2021 Taxonomy of Generative Models Direct GAN Generative models Explicit density Implicit density Markov Chain Tractable density Approximate density GSN Fully Visible Belief Nets - NADE - MADE Variational Markov Chain - PixelRNN/CNN Variational Autoencoder Boltzmann Machine - NICE / RealNVP - Glow Figure copyright and adapted from Ian Goodfellow, Tutorial on Generative Adversarial Networks, 2017. - Ffjord Fei-Fei Li, Ranjay Krishna, Danfei Xu Lecture 12 - 16 May 11, 2021 Taxonomy of Generative Models Direct Today: discuss 3 most GAN popular types of generative Generative models models today Explicit density Implicit density Markov Chain Tractable density Approximate density GSN Fully Visible Belief Nets - NADE - MADE Variational Markov Chain - PixelRNN/CNN Variational Autoencoder Boltzmann Machine - NICE / RealNVP - Glow Figure copyright and adapted from Ian Goodfellow, Tutorial on Generative Adversarial Networks, 2017. - Ffjord Fei-Fei Li, Ranjay Krishna, Danfei Xu Lecture 12 - 17 May 11, 2021 PixelRNN and PixelCNN (A very brief overview) Fei-Fei Li, Ranjay Krishna, Danfei Xu Lecture 12 - 18 May 11, 2021 Fully visible belief network (FVBN) Explicit density model Likelihood of Joint likelihood of each image x pixel in the image Fei-Fei Li, Ranjay Krishna, Danfei Xu Lecture 12 - 19 May 11, 2021 Fully visible belief network (FVBN) Explicit density model Use chain rule to decompose likelihood of an image x into product of 1-d distributions: Likelihood of Probability of i’th pixel value image x given all previous pixels Then maximize likelihood of training data Fei-Fei Li, Ranjay Krishna, Danfei Xu Lecture 12 - 20 May 11, 2021 Fully visible belief network (FVBN) Explicit density model Use chain rule to decompose likelihood of an image x into product of 1-d distributions: Likelihood of Probability of i’th pixel value image x given all previous pixels Complex distribution over pixel values => Express using a neural Then maximize likelihood of training data network! Fei-Fei Li, Ranjay Krishna, Danfei Xu Lecture 12 - 21 May 11, 2021 Recurrent Neural Network x2 x3 x4 xn h h h 0 1 2 h3 RNN RNN RNN ... RNN xn- x1 x2 x3 1 Fei-Fei Li, Ranjay Krishna, Danfei Xu Lecture 12 - 22 May 11, 2021 PixelRNN [van der Oord et al. 2016] Generate image pixels starting from corner Dependency on previous pixels modeled using an RNN (LSTM) Fei-Fei Li, Ranjay Krishna, Danfei Xu Lecture 12 - 23 May 11, 2021 PixelRNN [van der Oord et al. 2016] Generate image pixels starting from corner Dependency on previous pixels modeled using an RNN (LSTM) Fei-Fei Li, Ranjay Krishna, Danfei Xu Lecture 12 - 24 May 11, 2021 PixelRNN [van der Oord et al. 2016] Generate image pixels starting from corner Dependency on previous pixels modeled using an RNN (LSTM) Fei-Fei Li, Ranjay Krishna, Danfei Xu Lecture 12 - 25 May 11, 2021 PixelRNN [van der Oord et al. 2016] Generate image pixels starting from corner Dependency on previous pixels modeled using an RNN (LSTM) Drawback: sequential generation is slow in both training and inference! Fei-Fei Li, Ranjay Krishna, Danfei Xu Lecture 12 - 26 May 11, 2021 PixelCNN [van der Oord et al. 2016] Still generate image pixels starting from corner Dependency on previous pixels now modeled using a CNN over context region (masked convolution) Figure copyright van der Oord et al., 2016. Reproduced with permission. Fei-Fei Li, Ranjay Krishna, Danfei Xu Lecture 12 - 27 May 11, 2021 PixelCNN [van der Oord et al. 2016] Still generate image pixels starting from corner Dependency on previous pixels now modeled using a CNN over context region (masked convolution) Training is faster than PixelRNN (can parallelize convolutions since context region values known from training images) Generation is still slow: For a 32x32 image, we need to do forward passes of Figure copyright van der Oord et al., 2016. Reproduced with permission. the network 1024 times for a single image Fei-Fei Li, Ranjay Krishna, Danfei Xu Lecture 12 - 28 May 11, 2021 Generation Samples 32x32 CIFAR-10 32x32 ImageNet Figures copyright Aaron van der Oord et al., 2016. Reproduced with permission. Fei-Fei Li, Ranjay Krishna, Danfei Xu Lecture 12 - 29 May 11, 2021 PixelRNN

View Full Text

Details

  • File Type
    pdf
  • Upload Time
    -
  • Content Languages
    English
  • Upload User
    Anonymous/Not logged-in
  • File Pages
    132 Page
  • File Size
    -

Download

Channel Download Status
Express Download Enable

Copyright

We respect the copyrights and intellectual property rights of all users. All uploaded documents are either original works of the uploader or authorized works of the rightful owners.

  • Not to be reproduced or distributed without explicit permission.
  • Not used for commercial purposes outside of approved use cases.
  • Not used to infringe on the rights of the original creators.
  • If you believe any content infringes your copyright, please contact us immediately.

Support

For help with questions, suggestions, or problems, please contact us