Skip to content

Latest commit

 

History

History
41 lines (23 loc) · 1.6 KB

File metadata and controls

41 lines (23 loc) · 1.6 KB

Depth-Map-Prediction-from-a-Single-Image-using-a-Multi-Scale-Deep-Network

PyTorch implementation from the papers:
https://cs.nyu.edu/~deigen/depth/depth_nips14.pdf
https://arxiv.org/pdf/1411.4734v4.pdf

Extended model architecture and loss fn to the newer paper.

Model Results:
(image, ground truth depth, model prediction) alt text

Model Arch:

Loss Fn:

Predictions on test image:

Pros:

  • Can detect object boundaries well, due to added image gradient component in the newer loss fn.
  • Prediction quality is decent considering from single image

Cons:

  • Model produces depthmaps at lower resolution (320x240)
  • Depthmaps lack clarity
  • Model is really large, ~900MB, inference time is ~2s for a mini-batch of 8 (640x480) images

Model weights: https://oregonstate.box.com/s/p3lbkgiwufg9rxfgx53c4svnzz2lz9av
NYU Depth Datasets: https://cs.nyu.edu/~silberman/datasets/