Depth-Aware Panoptic Segmentation

Publikation: Beitrag in FachzeitschriftKonferenzaufsatz in FachzeitschriftForschungPeer-Review

Autorschaft

Forschungs-netzwerk anzeigen

Details

OriginalspracheEnglisch
Seiten (von - bis)153-161
Seitenumfang9
FachzeitschriftISPRS Annals of the Photogrammetry, Remote Sensing and Spatial Information Sciences
AusgabenummerX-2-2024
PublikationsstatusVeröffentlicht - 10 Juni 2024
Veranstaltung2024 ISPRS TC II Mid-term Symposium on The Role of Photogrammetry for a Sustainable World - Las Vegas, USA / Vereinigte Staaten
Dauer: 11 Juni 202414 Juni 2024

Abstract

Panoptic segmentation unifies semantic and instance segmentation and thus delivers a semantic class label and, for so-called thing classes, also an instance label per pixel. The differentiation of distinct objects of the same class with a similar appearance is particularly challenging and frequently causes such objects to be incorrectly assigned to a single instance. In the present work, we demonstrate that information on the 3D geometry of the observed scene can be used to mitigate this issue: We present a novel CNN-based method for panoptic segmentation which processes RGB images and depth maps given as input in separate network branches and fuses the resulting feature maps in a late fusion manner. Moreover, we propose a new depth-aware dice loss term which penalises the assignment of pixels to the same thing instance based on the difference between their associated distances to the camera. Experiments carried out on the Cityscapes dataset show that the proposed method reduces the number of objects that are erroneously merged into one thing instance and outperforms the method used as basis by +2.2% in terms of panoptic quality.

ASJC Scopus Sachgebiete

Zitieren

Depth-Aware Panoptic Segmentation. / Nguyen, Tuan; Mehltretter, Max; Rottensteiner, Franz.
in: ISPRS Annals of the Photogrammetry, Remote Sensing and Spatial Information Sciences, Nr. X-2-2024, 10.06.2024, S. 153-161.

Publikation: Beitrag in FachzeitschriftKonferenzaufsatz in FachzeitschriftForschungPeer-Review

Nguyen, T, Mehltretter, M & Rottensteiner, F 2024, 'Depth-Aware Panoptic Segmentation', ISPRS Annals of the Photogrammetry, Remote Sensing and Spatial Information Sciences, Nr. X-2-2024, S. 153-161. https://doi.org/10.48550/arXiv.2405.10947, https://doi.org/10.5194/isprs-annals-X-2-2024-153-2024
Nguyen, T., Mehltretter, M., & Rottensteiner, F. (2024). Depth-Aware Panoptic Segmentation. ISPRS Annals of the Photogrammetry, Remote Sensing and Spatial Information Sciences, (X-2-2024), 153-161. https://doi.org/10.48550/arXiv.2405.10947, https://doi.org/10.5194/isprs-annals-X-2-2024-153-2024
Nguyen T, Mehltretter M, Rottensteiner F. Depth-Aware Panoptic Segmentation. ISPRS Annals of the Photogrammetry, Remote Sensing and Spatial Information Sciences. 2024 Jun 10;(X-2-2024):153-161. doi: 10.48550/arXiv.2405.10947, 10.5194/isprs-annals-X-2-2024-153-2024
Nguyen, Tuan ; Mehltretter, Max ; Rottensteiner, Franz. / Depth-Aware Panoptic Segmentation. in: ISPRS Annals of the Photogrammetry, Remote Sensing and Spatial Information Sciences. 2024 ; Nr. X-2-2024. S. 153-161.
Download
@article{5dd09e087af54316a14ce00313e5e590,
title = "Depth-Aware Panoptic Segmentation",
abstract = "Panoptic segmentation unifies semantic and instance segmentation and thus delivers a semantic class label and, for so-called thing classes, also an instance label per pixel. The differentiation of distinct objects of the same class with a similar appearance is particularly challenging and frequently causes such objects to be incorrectly assigned to a single instance. In the present work, we demonstrate that information on the 3D geometry of the observed scene can be used to mitigate this issue: We present a novel CNN-based method for panoptic segmentation which processes RGB images and depth maps given as input in separate network branches and fuses the resulting feature maps in a late fusion manner. Moreover, we propose a new depth-aware dice loss term which penalises the assignment of pixels to the same thing instance based on the difference between their associated distances to the camera. Experiments carried out on the Cityscapes dataset show that the proposed method reduces the number of objects that are erroneously merged into one thing instance and outperforms the method used as basis by +2.2% in terms of panoptic quality.",
keywords = "Dice Loss, Panoptic Segmentation, RGB Depth Fusion",
author = "Tuan Nguyen and Max Mehltretter and Franz Rottensteiner",
note = "Publisher Copyright: {\textcopyright} Author(s) 2024.; 2024 ISPRS TC II Mid-term Symposium on The Role of Photogrammetry for a Sustainable World ; Conference date: 11-06-2024 Through 14-06-2024",
year = "2024",
month = jun,
day = "10",
doi = "10.48550/arXiv.2405.10947",
language = "English",
pages = "153--161",
number = "X-2-2024",

}

Download

TY - JOUR

T1 - Depth-Aware Panoptic Segmentation

AU - Nguyen, Tuan

AU - Mehltretter, Max

AU - Rottensteiner, Franz

N1 - Publisher Copyright: © Author(s) 2024.

PY - 2024/6/10

Y1 - 2024/6/10

N2 - Panoptic segmentation unifies semantic and instance segmentation and thus delivers a semantic class label and, for so-called thing classes, also an instance label per pixel. The differentiation of distinct objects of the same class with a similar appearance is particularly challenging and frequently causes such objects to be incorrectly assigned to a single instance. In the present work, we demonstrate that information on the 3D geometry of the observed scene can be used to mitigate this issue: We present a novel CNN-based method for panoptic segmentation which processes RGB images and depth maps given as input in separate network branches and fuses the resulting feature maps in a late fusion manner. Moreover, we propose a new depth-aware dice loss term which penalises the assignment of pixels to the same thing instance based on the difference between their associated distances to the camera. Experiments carried out on the Cityscapes dataset show that the proposed method reduces the number of objects that are erroneously merged into one thing instance and outperforms the method used as basis by +2.2% in terms of panoptic quality.

AB - Panoptic segmentation unifies semantic and instance segmentation and thus delivers a semantic class label and, for so-called thing classes, also an instance label per pixel. The differentiation of distinct objects of the same class with a similar appearance is particularly challenging and frequently causes such objects to be incorrectly assigned to a single instance. In the present work, we demonstrate that information on the 3D geometry of the observed scene can be used to mitigate this issue: We present a novel CNN-based method for panoptic segmentation which processes RGB images and depth maps given as input in separate network branches and fuses the resulting feature maps in a late fusion manner. Moreover, we propose a new depth-aware dice loss term which penalises the assignment of pixels to the same thing instance based on the difference between their associated distances to the camera. Experiments carried out on the Cityscapes dataset show that the proposed method reduces the number of objects that are erroneously merged into one thing instance and outperforms the method used as basis by +2.2% in terms of panoptic quality.

KW - Dice Loss

KW - Panoptic Segmentation

KW - RGB Depth Fusion

UR - http://www.scopus.com/inward/record.url?scp=85199892117&partnerID=8YFLogxK

U2 - 10.48550/arXiv.2405.10947

DO - 10.48550/arXiv.2405.10947

M3 - Conference article

AN - SCOPUS:85199892117

SP - 153

EP - 161

JO - ISPRS Annals of the Photogrammetry, Remote Sensing and Spatial Information Sciences

JF - ISPRS Annals of the Photogrammetry, Remote Sensing and Spatial Information Sciences

SN - 2194-9042

IS - X-2-2024

T2 - 2024 ISPRS TC II Mid-term Symposium on The Role of Photogrammetry for a Sustainable World

Y2 - 11 June 2024 through 14 June 2024

ER -

Von denselben Autoren