AmbientGAN: Generative Models from Lossy Measurements - Deepest-Project/Greedy-Survey GitHub Wiki

Resources

Abstract

๋ฌธ์ œ์˜์‹

Current techniques for training generative models require access to fully-observed samples.
It is expensive or even impossible to obtain fully-observed samples, but economical to obtain partial, noisy observations.
We show that the true underlying distribution can be provably recovered even in the presence of per-sample information loss for a class of measurement models.
We call this AmbientGAN.

1. Introduction

  • GAN์— ๋Œ€ํ•œ ์ผ๋ฐ˜์ ์ธ ์„ค๋ช…์ด ๋‚˜์˜ต๋‹ˆ๋‹ค.
  • train์„ ์œ„ํ•ด์„œ๋Š” large number of fully-observed samples ์ด ํ•„์š”ํ•˜๋‹ค.
  • ์–ด๋–ค ๋ถ„์•ผ์—์„œ๋Š” samples์˜ ํš๋“ ๋น„์šฉ์ด ๋„ˆ๋ฌด ๋†’๊ฑฐ๋‚˜ ์‹ค์šฉ์ ์ด์ง€ ์•Š์€ ๊ฒฝ์šฐ๊ฐ€ ์žˆ๋‹ค.(many sensing and tomography problems, MRI, CT Scan)
  • ๊ทธ๋Ÿฐ๋ฐ ์ฒ˜์Œ๋ถ€ํ„ฐ sensing์ด ๋น„์‹ธ๋‹ค๋ฉด? ์–ด๋–ป๊ฒŒ ํ›ˆ๋ จ์ด ๊ฐ€๋Šฅํ•  ์ •๋„๋กœ ๋ฐ์ดํ„ฐ๋ฅผ ๋ชจ์„ ์ˆ˜ ์žˆ์„ ๊ฒƒ์ธ๊ฐ€?

This work solves this chicken-and-egg problem by training a generative model directly from noisy or incomplete samples.
We show that our observations can be even projections or more general measurements of different types and the unknown distribution is still provably recoverable. A critical assumption for our framework and theory to work is that the measurement process is known and satisfies certain technical conditions.

  • The idea is simple: rather than distinguish a real image from a generated image as in a traditional GAN, our discriminator must distinguish a real measurement from a simulated measurement of a generated image.

2. Related Work

  • Our work is also closely related to [Gadelha et al. (2016)] where the authors create 3D object shapes from a dataset of 2D projections. We note that their setup is a special case of the AmbientGAN framework where the measurement process creates 2D projections using weighted sums of voxel occupancies.

3. Notation and our approach


* ์ถœ์ฒ˜: https://ml-dnn.tistory.com/9#2-1-Notation


4. Measurement Models

  • Block-Pixels: ํ™•๋ฅ  p์— ๋”ฐ๋ผ ๋…๋ฆฝ์ ์œผ๋กœ ๊ฐ๊ฐ์˜ ํ”ฝ์…€ ๊ฐ’์ด 0
  • Convolve+Noise: k๋ฅผ convolution kernel ์ด๋ผ๊ณ  ํ•˜๊ณ , ์ด๋ผ๊ณ  ํ•˜๋ฉด,
     measurements๋Š”
        
  • Block-Patch: ์ž„์˜๋กœ ์„ ์ •๋œ k x k patch๊ฐ€ 0์œผ๋กœ ์„ค์ •.
  • Keep-Patch: ์ž„์˜๋กœ ์„ ์ •๋œ k x k patch ๋ฐ”๊นฅ์˜ ๊ฐ’๋“ค์ด 0์œผ๋กœ ์„ค์ •.
  • Extract-Patch: ์ž„์˜๋กœ ์„ ์ •๋œ k x k patch๊ฐ€ ์ถ”์ถœ๋จ.
  • Pad-Rotate-Project: 4๋ฉด์„ ๋ชจ๋‘ ํŒจ๋”ฉํ•˜๊ณ  ํšŒ์ „์„ ์‹œํ‚ด.
  • Pad-Rotate-Project-์„ธํƒ€: ์œ„์™€ ๊ฐ™์€๋ฐ ๋Œ€์‹  ํšŒ์ „๊ฐ์„ ์ •ํ•ด์คŒ.
  • Gaussian-Projection: random Gaussian vector๋กœ projectํ•จ.

5. Theoretical Result


6. Datasets and Model Architectures

  • MNIST, CelebA, CIFAR-10
  • MNIST:
      first model- conditional DCGAN, second model- unconditional Wasserstein GAN with gradient penalty.
  • celebA:
      unconditional DCGAN
  • CIFAR-10:
      Auxiliary Classifier Wasserstein GAN with gradient penalty(ACWGANGP)

7. BASELINES

  • "ignore" baseline: A crude baseline is to ignore that any measurement happened at all.
  • "stronger" baseline: If the measurement functions were invertible, and we observed for each measurement y_i in our dataset, we could just invert the functions to objain full-samples
     
  1. We may not observe
  2. Functions may not be invertible.

8. Qualitative Results

์ž์„ธํ•œ ์„ค๋ช…์€ ๋…ผ๋ฌธ์œผ๋กœ ๋Œ€์ฒดํ•˜๊ฒ ์Šต๋‹ˆ๋‹ค.

9. Quantitative Results

์ž์„ธํ•œ ์„ค๋ช…์€ ๋…ผ๋ฌธ์œผ๋กœ ๋Œ€์ฒดํ•˜๊ฒ ์Šต๋‹ˆ๋‹ค.

โš ๏ธ **GitHub.com Fallback** โš ๏ธ