MFGNet: Dynamic Modality-Aware Filter Generation for RGB-T Tracking | IEEE Journals & Magazine | IEEE Xplore

MFGNet: Dynamic Modality-Aware Filter Generation for RGB-T Tracking


Abstract:

Many RGB-T trackers attempt to attain robust feature representation by utilizing an adaptive weighting scheme (or attention mechanism). Different from these works, we pro...Show More

Abstract:

Many RGB-T trackers attempt to attain robust feature representation by utilizing an adaptive weighting scheme (or attention mechanism). Different from these works, we propose a new dynamic modality-aware filter generation module (named MFGNet) to boost the message communication between visible and thermal data by adaptively adjusting the convolutional kernels for various input images in practical tracking. Given the image pairs as input, we first encode their features with the backbone network. Then, we concatenate these feature maps and generate dynamic modality-aware filters with two independent networks. The visible and thermal filters will be used to conduct a dynamic convolutional operation on their corresponding input feature maps respectively. Inspired by residual connection, both the generated visible and thermal feature maps will be summarized with input feature maps. The augmented feature maps will be fed into the RoI align module to generate instance-level features for subsequent classification. To address issues caused by heavy occlusion, fast motion and out-of-view, we propose to conduct a joint local and global search by exploiting a new direction-aware target driven attention mechanism. The spatial and temporal recurrent neural network is used to capture the direction-aware context for accurate global attention prediction. Extensive experiments on three large-scale RGB-T tracking benchmark datasets validated the effectiveness of our proposed algorithm.
Published in: IEEE Transactions on Multimedia ( Volume: 25)
Page(s): 4335 - 4348
Date of Publication: 11 May 2022

ISSN Information:

Funding Agency:


I. Introduction

Object tracking is a popular research topic in computer vision that aims to locate a determined object (initialized in the first frame) in each video frame. It has been widely used in many applications, such as intelligent surveillance, automatic driving, and unmanned aerial vehicles. Although it has already achieved great success in recent years with robust target representation brought by deep neural network [1]–[15], these trackers still suffer from challenging factors, e.g., illumination, scale variation, and fast motion.

Contact IEEE to Subscribe

References

References is not available for this document.