1603.08152 - hassony2/inria-research-wiki GitHub Wiki

GMDL (Workshop at ECCV) 2016

[arxiv 1603.08152] How useful is photo-realistic rendering for visual learning? [PDF] [notes]

Yair Movshovitz-Attias, Takeo Kanade, Yaser Sheikh

read 17/04/2018

Objective

Assess effect of realistic lighting and rendering on generalization to real data Show that realism is important towards object viewpoint estimation

Synthesis

Dataset generation

Rendering settings

  • Randomized light:

  • temperature

  • intensity

  • elevation on a sphere

  • Randomized Shutter speed and F-stop

  • Lens vignetting

  • Backgrounds from VOC where no car is present

  • All images are JPG compressed

Training augmentation

  • Color jittering
  • Color channel swapping
  • Image degradation by downsampling
  • place rectangular patches from uniform color or Pascal dataset on the renders

Train-test split

819.000 images for training, 1.800 for testing (leave one car model out)

Experiments

Test datasets

  • CMU-Car (camera matrices for 3.240 cars, manual annotations of landmark points to determine view point)
  • PASCAL3D+

Viewpoint prediction

  • Two synthetic datasets:
    • RenderScene : one car model shot from 1800 angles
    • RenderCar: use entire set of 3D CAD vehicles models
  • Use ground truth bounding boxes Using RenderCar is on par or outperforms training on small real datasets alone, combining leads best results

Results

Lighting and Material complexity

Complex Material and Directional lighting > Complex material, ambient lighting > Simple material and ambient lighting

Mixture of synthetic and real

At given train size, Mixing 25% to 50% synthetic with real performs best, slightly above (1% absolute diff in angular error) using only real and significantly better then using only synthetic (8% absolute diff in angular error)