Real time computation of depth from defocus

Abstract

A new range sensing method based on depth from defocus is described It uses illumination pattern projection to give texture to the object surface Then the image of the scene is split into two images with di erent focus settings and sensed simultaneously The contrast map of the two images are computed and compared pixel by pixel to produce a dense depth map The illumination pattern and the focus operator to extract the contrast map are designed to achieve nest spatial resolution of the computed depth map and to maximize response of the focus operator As the algorithm uses only local operations such as convolution and lookup table the depth map can be computed rapidly on a data ow image processing hardware As this projects an illumination pattern and detects the two images with di erent focus setting from exactly the same direction it does not share the problem of shadowing and occlusion with triangulation based method and stereo It s speed and accuracy are demonstrated using a prototype system The prototype generates range maps at frame sec with a depth resolution of relative to the object distance The proposed sensor is composed of o the shelf compo nents and outperforms commercial range sensors through its ability to produce complete three dimensional shape information at video rate

Extracted Key Phrases

2 Figures and Tables

Cite this paper

@inproceedings{Watanabe1995RealTC, title={Real time computation of depth from defocus}, author={Masahiro Watanabe and Shree K Nayary and Minori Noguchi}, year={1995} }