SUN3D: A Database of Big Spaces Reconstructed using SfM and Object Labels
International Conference on Computer Vision (ICCV), December 2013
Abstract
Existing scene understanding datasets contain only a limited set of views of a place, and they lack representations of complete 3D spaces. In this paper, we introduce SUN3D, a large-scale RGB-D video database with camera pose and object labels, capturing the full 3D extent of many places. The tasks that go into constructing such a dataset are difficult in isolation – hand-labeling videos is painstaking, and structure from motion (SfM) is unreliable for large spaces. But if we combine them together, we make the dataset construction task much easier. First, we introduce an intuitive labeling tool that uses a partial reconstruction to propagate labels from one frame to another. Then we use the object labels to fix errors in the reconstruction. For this, we introduce a generalization of bundle adjustment that incorporates object-to-object correspondences. This algorithm works by constraining points for the same object from different frames to lie inside a fixed-size bounding box, parameterized by its rotation and translation. The SUN3D database, the source code for the generalized bundle adjustment, and the web-based 3D annotation tool are all available at http://sun3d.cs.princeton.edu.
Citation
Jianxiong Xiao, Andrew Owens, and Antonio Torralba.
"SUN3D: A Database of Big Spaces Reconstructed using SfM and Object Labels."
International Conference on Computer Vision (ICCV), December 2013.
BibTeX
@inproceedings{Xiao:2013:SAD, author = "Jianxiong Xiao and Andrew Owens and Antonio Torralba", title = "{SUN3D}: A Database of Big Spaces Reconstructed using {SfM} and Object Labels", booktitle = "International Conference on Computer Vision (ICCV)", year = "2013", month = dec }