Loading [MathJax]/extensions/MathMenu.js
Modality-Induced Transfer-Fusion Network for RGB-D and RGB-T Salient Object Detection | IEEE Journals & Magazine | IEEE Xplore

Modality-Induced Transfer-Fusion Network for RGB-D and RGB-T Salient Object Detection


Abstract:

The ability of capturing the complementary information of multi-modality data is critical to the development of multi-modality salient object detection (SOD). Most of exi...Show More

Abstract:

The ability of capturing the complementary information of multi-modality data is critical to the development of multi-modality salient object detection (SOD). Most of existing studies attempt to integrate multi-modality information through various fusion strategies. However, most of these methods ignore the inherent differences in multi-modality data, resulting in poor performance when dealing with some challenging scenarios. In this paper, we propose a novel Modality-Induced Transfer-Fusion Network (MITF-Net) for RGB-D and RGB-T SOD by fully exploring the complementarity in multi-modality data. Specifically, we first deploy a modality transfer fusion (MTF) module to bridge the semantic gap between single and multi-modality data, and then mine the cross-modality complementarity based on point-to-point structural similarity information. Then, we design a cycle-separated attention (CSA) module to optimize the cross-layer information recurrently, and measure the effectiveness of cross-layer features through point-wise convolution-based multi-scale channel attention. Furthermore, we refine the boundaries in the decoding stage to obtain high-quality saliency maps with sharp boundaries. Extensive experiments on 13 RGB-D and RGB-T SOD datasets show that the proposed MITF-Net achieves a competitive and excellent performance.
Page(s): 1787 - 1801
Date of Publication: 19 October 2022

ISSN Information:

Funding Agency:


Contact IEEE to Subscribe

References

References is not available for this document.