PASS: Panoramic Annular Semantic Segmentation
Autores
Yang, Kailun; Hu, Xinxin; Bergasa Pascual, Luis MiguelIdentificadores
Enlace permanente (URI): http://hdl.handle.net/10017/42859DOI: 10.1109/TITS.2019.2938965
ISSN: 1524-9050
Editor
IEEE
Fecha de publicación
2019-09-12Fecha fin de embargo
2020-09-12Patrocinadores
Ministerio de Economía y Competitividad
Comunidad de Madrid
Cita bibliográfica
K. Yang, X. Hu, L. M. Bergasa, E. Romera and K. Wang, 2019, "PASS: Panoramic Annular Semantic Segmentation," IEEE Transactions on Intelligent Transportation Systems, doi: 10.1109/TITS.2019.2938965
Palabras clave
Semantics
Cameras
Image segmentation
Sensors
Navigation
Task analysis
Benchmark testing
Intelligent vehicles
Scene parsing
Semantic segmentation
Scene understanding
Panoramic annular images
Proyectos
info:eu-repo/grantAgreement/MINECO//TRA2015-70501-C2-1-R/ES/VEHICULO INTELIGENTE PARA PERSONAS MAYORES/
info:eu-repo/grantAgreement/CAM//P2018%2FNMT-4331/ES/Madrid Robotics Digital Innovation Hub/RoboCity2030-DIH-CM
Tipo de documento
info:eu-repo/semantics/article
Versión
info:eu-repo/semantics/acceptedVersion
Versión del editor
https://doi.org/10.1109/TITS.2019.2938965Derechos
Attribution-NonCommercial-NoDerivatives 4.0 Internacional
© 2019 IEEE
Derechos de acceso
info:eu-repo/semantics/openAccess
Resumen
Pixel-wise semantic segmentation is capable of unifying most of driving scene perception tasks, and has enabled striking progress in the context of navigation assistance, where an entire surrounding sensing is vital. However, current mainstream semantic segmenters are predominantly benchmarked against datasets featuring narrow Field of View (FoV), and a large part of vision-based intelligent vehicles use only a forward-facing camera. In this paper, we propose a Panoramic Annular Semantic Segmentation (PASS) framework to perceive the whole surrounding based on a compact panoramic annular lens system and an online panorama unfolding process. To facilitate the training of PASS models, we leverage conventional FoV imaging datasets, bypassing the efforts entailed to create fully dense panoramic annotations. To consistently exploit the rich contextual cues in the unfolded panorama, we adapt our real-time ERF-PSPNet to predict semantically meaningful feature maps in different segments, and fuse them to fulfill panoramic scene parsing. The innovation lies in the network adaptation to enable smooth and seamless segmentation, combined with an extended set of heterogeneous data augmentations to attain robustness in panoramic imagery. A comprehensive variety of experiments demonstrates the effectiveness for real-world surrounding perception in a single PASS, while the adaptation proposal is exceptionally positive for state-of-the-art efficient networks.
Ficheros en el ítem
Ficheros | Tamaño | Formato |
|
---|---|---|---|
PASS_Yang_IEEE_TITS_2019.pdf | 945.2Kb |
![]() |
Ficheros | Tamaño | Formato |
|
---|---|---|---|
PASS_Yang_IEEE_TITS_2019.pdf | 945.2Kb |
![]() |
Colecciones
- ELECTRON - Artículos [152]
- ROBESAFE - Artículos [37]