Learning Deep Features for Scene Recognition using Places Database
Advances in Neural Information Processing Systems, December 2014
Abstract
Scene recognition is one of the hallmark tasks of computer vision, allowing definition
of a context for object recognition. Whereas the tremendous recent progress
in object recognition tasks is due to the availability of large datasets like ImageNet
and the rise of Convolutional Neural Networks (CNNs) for learning high-level features,
performance at scene recognition has not attained the same level of success.
This may be because current deep features trained from ImageNet are not competitive
enough for such tasks. Here, we introduce a new scene-centric database called
Places with over 7 million labeled pictures of scenes. We propose new methods
to compare the density and diversity of image datasets and show that Places is as
dense as other scene datasets and has more diversity. Using CNN, we learn deep
features for scene recognition tasks, and establish new state-of-the-art results on
several scene-centric datasets. A visualization of the CNN layers’ responses allows
us to show differences in the internal representations of object-centric and
scene-centric networks.
Citation
Bolei Zhou, Agata Lapedriza, Jianxiong Xiao, Antonio Torralba, and Aude Oliva.
"Learning Deep Features for Scene Recognition using Places Database."
Advances in Neural Information Processing Systems, December 2014.
BibTeX
@inproceedings{Zhou:2014:LDF, author = "Bolei Zhou and Agata Lapedriza and Jianxiong Xiao and Antonio Torralba and Aude Oliva", title = "Learning Deep Features for Scene Recognition using Places Database", booktitle = "Advances in Neural Information Processing Systems", year = "2014", month = dec }