Abstract:
Semantic segmentation of images promises numerous benefits for augmented reality applications. However, in such applications typical scenes are challenging for current se...Show MoreMetadata
Abstract:
Semantic segmentation of images promises numerous benefits for augmented reality applications. However, in such applications typical scenes are challenging for current segmentation algorithms due to high variability in object appearances and distribution. We propose a new cascaded loss fusion strategy to improve the training schedule of state-of-the-art realtime RGB-D semantic segmentation architectures. We employ methods developed in the context of multi-task learning to solve the multiclass and multi-loss learning problems in semantic segmentation. Through our quantitative evaluation on the NYUv2 [3] and SUNRGB-D [4] benchmark datasets, we show improvement over the state-of-the-art approaches. Furthermore, our approach improves results qualitatively on both the benchmark datasets as well as on our own recordings of some scenarios that are typical for head-mounted cameras.
Published in: 2024 IEEE International Conference on Artificial Intelligence and eXtended and Virtual Reality (AIxVR)
Date of Conference: 17-19 January 2024
Date Added to IEEE Xplore: 28 February 2024
ISBN Information: