PASS: Panoramic Annular Semantic Segmentation
Authors
Yang, Kailun; Hu, Xinxin; Bergasa Pascual, Luis MiguelIdentifiers
Permanent link (URI): http://hdl.handle.net/10017/42859DOI: 10.1109/TITS.2019.2938965
ISSN: 1524-9050
Publisher
IEEE
Date
2019-09-12Embargo end date
2020-09-12Funders
Ministerio de Economía y Competitividad
Comunidad de Madrid
Bibliographic citation
K. Yang, X. Hu, L. M. Bergasa, E. Romera and K. Wang, 2019, "PASS: Panoramic Annular Semantic Segmentation," IEEE Transactions on Intelligent Transportation Systems, doi: 10.1109/TITS.2019.2938965
Keywords
Semantics
Cameras
Image segmentation
Sensors
Navigation
Task analysis
Benchmark testing
Intelligent vehicles
Scene parsing
Semantic segmentation
Scene understanding
Panoramic annular images
Project
info:eu-repo/grantAgreement/MINECO//TRA2015-70501-C2-1-R/ES/VEHICULO INTELIGENTE PARA PERSONAS MAYORES/
info:eu-repo/grantAgreement/CAM//P2018%2FNMT-4331/ES/Madrid Robotics Digital Innovation Hub/RoboCity2030-DIH-CM
Document type
info:eu-repo/semantics/article
Version
info:eu-repo/semantics/acceptedVersion
Publisher's version
https://doi.org/10.1109/TITS.2019.2938965Rights
Attribution-NonCommercial-NoDerivatives 4.0 Internacional
© 2019 IEEE
Access rights
info:eu-repo/semantics/openAccess
Abstract
Pixel-wise semantic segmentation is capable of unifying most of driving scene perception tasks, and has enabled striking progress in the context of navigation assistance, where an entire surrounding sensing is vital. However, current mainstream semantic segmenters are predominantly benchmarked against datasets featuring narrow Field of View (FoV), and a large part of vision-based intelligent vehicles use only a forward-facing camera. In this paper, we propose a Panoramic Annular Semantic Segmentation (PASS) framework to perceive the whole surrounding based on a compact panoramic annular lens system and an online panorama unfolding process. To facilitate the training of PASS models, we leverage conventional FoV imaging datasets, bypassing the efforts entailed to create fully dense panoramic annotations. To consistently exploit the rich contextual cues in the unfolded panorama, we adapt our real-time ERF-PSPNet to predict semantically meaningful feature maps in different segments, and fuse them to fulfill panoramic scene parsing. The innovation lies in the network adaptation to enable smooth and seamless segmentation, combined with an extended set of heterogeneous data augmentations to attain robustness in panoramic imagery. A comprehensive variety of experiments demonstrates the effectiveness for real-world surrounding perception in a single PASS, while the adaptation proposal is exceptionally positive for state-of-the-art efficient networks.
Files in this item
Files | Size | Format |
|
---|---|---|---|
PASS_Yang_IEEE_TITS_2019.pdf | 945.2Kb |
![]() |
Files | Size | Format |
|
---|---|---|---|
PASS_Yang_IEEE_TITS_2019.pdf | 945.2Kb |
![]() |
Collections
- ELECTRON - Artículos [152]
- ROBESAFE - Artículos [37]