To refer to this page use:
|Abstract:||The depth information of RGB-D sensors has greatly simplified some common challenges in computer vision and enabled breakthroughs for several tasks. In this paper, we propose to use depth maps for object detection and design a 3D detector to overcome the major difficulties for recognition, namely the variations of texture, illumination, shape, viewpoint, clutter, occlusion, self-occlusion and sensor noises. We take a collection of 3D CAD models and render each CAD model from hundreds of viewpoints to obtain synthetic depth maps. For each depth rendering, we extract features from the 3D point cloud and train an Exemplar-SVM classifier. During testing and hard-negative mining, we slide a 3D detection window in 3D space. Experiment results show that our 3D detector significantly outperforms the state-of-the-art algorithms for both RGB and RGB-D images, and achieves about ×1.7 improvement on average precision compared to DPM and R-CNN. All source code and data are available online.|
|Citation:||Song, Shuran, and Jianxiong Xiao. "Sliding Shapes for 3D Object Detection in Depth Images." In European Conference on Computer Vision (2014): pp. 634-651. doi:10.1007/978-3-319-10599-4_41|
|Pages:||634 - 651|
|Type of Material:||Conference Article|
|Journal/Proceeding Title:||European Conference on Computer Vision|
Items in OAR@Princeton are protected by copyright, with all rights reserved, unless otherwise indicated.