日本語
 
Help Privacy Policy ポリシー/免責事項
  詳細検索ブラウズ

アイテム詳細


公開

成果報告書

CityPersons: A Diverse Dataset for Pedestrian Detection

MPS-Authors
/persons/resource/persons134279

Zhang,  Shanshan
Computer Vision and Multimodal Computing, MPI for Informatics, Max Planck Society;

/persons/resource/persons79212

Benenson,  Rodrigo
Computer Vision and Multimodal Computing, MPI for Informatics, Max Planck Society;

/persons/resource/persons45383

Schiele,  Bernt
Computer Vision and Multimodal Computing, MPI for Informatics, Max Planck Society;

External Resource
There are no locators available
Fulltext (restricted access)
There are currently no full texts shared for your IP range.
フルテキスト (公開)

arXiv:1702.05693.pdf
(プレプリント), 9MB

付随資料 (公開)
There is no public supplementary material available
引用

Zhang, S., Benenson, R., & Schiele, B. (2017). CityPersons: A Diverse Dataset for Pedestrian Detection. Retrieved from http://arxiv.org/abs/1702.05693.


引用: https://hdl.handle.net/11858/00-001M-0000-002D-7CA8-E
要旨
Convnets have enabled significant progress in pedestrian detection recently, but there are still open questions regarding suitable architectures and training data. We revisit CNN design and point out key adaptations, enabling plain FasterRCNN to obtain state-of-the-art results on the Caltech dataset. To achieve further improvement from more and better data, we introduce CityPersons, a new set of person annotations on top of the Cityscapes dataset. The diversity of CityPersons allows us for the first time to train one single CNN model that generalizes well over multiple benchmarks. Moreover, with additional training with CityPersons, we obtain top results using FasterRCNN on Caltech, improving especially for more difficult cases (heavy occlusion and small scale) and providing higher localization quality.