Corpus ID: 226237414

Unsupervised Monocular Depth Learning with Integrated Intrinsics and Spatio-Temporal Constraints

@article{Chen2020UnsupervisedMD,
  title={Unsupervised Monocular Depth Learning with Integrated Intrinsics and Spatio-Temporal Constraints},
  author={Kenny Chen and Alexandra Pogue and Brett Thomas Lopez and Ali-akbar Agha-mohammadi and Ankur M. Mehta},
  journal={ArXiv},
  year={2020},
  volume={abs/2011.01354}
}
Monocular depth inference has gained tremendous attention from researchers in recent years and remains as a promising replacement for expensive time-of-flight sensors, but issues with scale acquisition and implementation overhead still plague these systems. To this end, this work presents an unsupervised learning framework that is able to predict at-scale depth maps and egomotion, in addition to camera intrinsics, from a sequence of monocular images via a single network. Our method incorporates… Expand

Figures and Tables from this paper

References

SHOWING 1-10 OF 55 REFERENCES
Unsupervised Scale-consistent Depth and Ego-motion Learning from Monocular Video
TLDR
This paper proposes a geometry consistency loss for scale-consistent predictions and an induced self-discovered mask for handling moving objects and occlusions and is the first work to show that deep networks trained using unlabelled monocular videos can predict globally scale- Consistent camera trajectories over a long video sequence. Expand
Towards Real-Time Unsupervised Monocular Depth Estimation on CPU
TLDR
This paper proposes a novel architecture capable to quickly infer an accurate depth map on a CPU, even of an embedded system, using a pyramid of features extracted from a single input image, and is the first method enabling such performance on CPUs paving the way for effective deployment of unsupervised monocular depth estimation even on embedded systems. Expand
Unsupervised Learning of Depth and Ego-Motion from Video
TLDR
Empirical evaluation demonstrates the effectiveness of the unsupervised learning framework for monocular depth performs comparably with supervised methods that use either ground-truth pose or depth for training, and pose estimation performs favorably compared to established SLAM systems under comparable input settings. Expand
GeoNet: Unsupervised Learning of Dense Depth, Optical Flow and Camera Pose
TLDR
An adaptive geometric consistency loss is proposed to increase robustness towards outliers and non-Lambertian regions, which resolves occlusions and texture ambiguities effectively and achieves state-of-the-art results in all of the three tasks, performing better than previously unsupervised methods and comparably with supervised ones. Expand
Digging Into Self-Supervised Monocular Depth Estimation
TLDR
It is shown that a surprisingly simple model, and associated design choices, lead to superior predictions, and together result in both quantitatively and qualitatively improved depth maps compared to competing self-supervised methods. Expand
UnDEMoN: Unsupervised Deep Network for Depth and Ego-Motion Estimation
TLDR
A deep network based unsupervised visual odometry system for 6-DoF camera pose estimation and finding dense depth map for its monocular view is presented and is shown to provide superior performance in depth and ego-motion estimation compared to the existing state-of-the-art methods. Expand
Depth From Videos in the Wild: Unsupervised Monocular Depth Learning From Unknown Cameras
TLDR
This work is the first to learn the camera intrinsic parameters, including lens distortion, from video in an unsupervised manner, thereby allowing us to extract accurate depth and motion from arbitrary videos of unknown origin at scale. Expand
3D Packing for Self-Supervised Monocular Depth Estimation
TLDR
This work proposes a novel self-supervised monocular depth estimation method combining geometry with a new deep network, PackNet, learned only from unlabeled monocular videos, which outperforms other self, semi, and fully supervised methods on the KITTI benchmark. Expand
Unsupervised Learning of Monocular Depth Estimation and Visual Odometry with Deep Feature Reconstruction
TLDR
The use of stereo sequences for learning depth and visual odometry enables the use of both spatial and temporal photometric warp error, and constrains the scene depth and camera motion to be in a common, real-world scale. Expand
Unsupervised Monocular Depth Estimation with Left-Right Consistency
TLDR
This paper proposes a novel training objective that enables the convolutional neural network to learn to perform single image depth estimation, despite the absence of ground truth depth data, and produces state of the art results for monocular depth estimation on the KITTI driving dataset. Expand
...
1
2
3
4
5
...