ScanNet: Richly-Annotated 3D Reconstructions of Indoor Scenes
Author(s): Dai, Angela; Chang, Angel X; Savva, Manolis; Halber, Maciej; Funkhouser, Thomas; et al
DownloadTo refer to this page use:
http://arks.princeton.edu/ark:/88435/pr1k848
Abstract: | A key requirement for leveraging supervised deep learning methods is the availability of large, labeled datasets. Unfortunately, in the context of RGB-D scene understanding, very little data is available - current datasets cover a small range of scene views and have limited semantic annotations. To address this issue, we introduce ScanNet, an RGB-D video dataset containing 2.5M views in 1513 scenes annotated with 3D camera poses, surface reconstructions, and semantic segmentations. To collect this data, we designed an easy-to-use and scalable RGB-D capture system that includes automated surface reconstruction and crowd-sourced semantic annotation.We show that using this data helps achieve state-of-the-art performance on several 3D scene understanding tasks, including 3D object classification, semantic voxel labeling, and CAD model retrieval. |
Publication Date: | 2017 |
Citation: | Dai, Angela, Angel X. Chang, Manolis Savva, Maciej Halber, Thomas Funkhouser, and Matthias Niessner. "ScanNet: Richly-Annotated 3D Reconstructions of Indoor Scenes." In IEEE Conference on Computer Vision and Pattern Recognition (CVPR) (2017): pp. 2432-2443. doi:10.1109/CVPR.2017.261 |
DOI: | 10.1109/CVPR.2017.261 |
ISSN: | 1063-6919 |
Pages: | 2432 - 2443 |
Type of Material: | Conference Article |
Journal/Proceeding Title: | IEEE Conference on Computer Vision and Pattern Recognition |
Version: | Author's manuscript |
Items in OAR@Princeton are protected by copyright, with all rights reserved, unless otherwise indicated.