Loading web-font TeX/Math/Italic
Hierarchical Deep Multitask Learning With the Attention Mechanism for Similarity Learning | IEEE Journals & Magazine | IEEE Xplore

Hierarchical Deep Multitask Learning With the Attention Mechanism for Similarity Learning


Abstract:

Similarity learning is often adopted as an auxiliary task of deep multitask learning methods to learn discriminant features. Most existing approaches only use the single-...Show More

Abstract:

Similarity learning is often adopted as an auxiliary task of deep multitask learning methods to learn discriminant features. Most existing approaches only use the single-layer features extracted by the last fully connected layer, which ignores the abundant information of feature channels in lower layers. Besides, small cliques are the most commonly used methods in similarity learning tasks to model the correlation of data, which can lead to the limited relation learning. In this article, we present an end-to-end hierarchical deep multitask learning framework for similarity learning which can learn more discriminant features by sharing information from different layers of network and dealing with complex correlation. Its main task is graph similarity inference. We build focus graphs for each sample. Then, an attention mechanism and a node feature enhancing model are introduced into backbone network to extract the abundant and important channel information from multiple layers of network. In the similarity inference task, a relation enhancing mechanism is applied to graph convolutional network to leverage the crucial relation in channels, which can effectively facilitate the learning ability of the whole framework. The extensive experiments have been conducted to demonstrate the effectiveness of the proposed method on person reidentification and face clustering applications.
Published in: IEEE Transactions on Cognitive and Developmental Systems ( Volume: 14, Issue: 4, December 2022)
Page(s): 1729 - 1742
Date of Publication: 21 December 2021

ISSN Information:

Funding Agency:


I. Introduction

Deep metric learning (DML) as one of the similarity learning methods has attracted more and more attention recently in the field of deep learning [1]–[7]. Its commonly strategy is to exploit a deep end-to-end feature representation. DML methods take the relations between samples into consideration and map samples into a new embedding space where samples with the same label are closer while the samples with different labels are far apart. However, the features learned by DML methods may yield suboptimal results if they only modeled simple relations of the data. For this reason, many deep learning methods introduce a multitask learning mechanism [8], [9], i.e., optimizing the classification task and similarity learning task at the same time. In this way, these methods can achieve better results. However, there are still two problems existed in these methods.

A variety of existing methods only extract single-layer features from the last fully connected layer of deep neural networks [10]–[12]. As a matter of fact, the features extracted from images in low layers have abundant details like position, while high-layer features contain semantic information, such as shapes and targets. Hence, single-layer features may be sensitive to variations, such as viewpoints and illumination.

Most deep multitask metric learning methods organize the training samples into small cliques to compute their correlations, such as pairs [13], [14], triplets [15]–[18], and quadruplets [10]. Accordingly, the learned features may be discriminative only in cliques while not in the whole embedding space due to limited correlation.

Contact IEEE to Subscribe

References

References is not available for this document.