Abstract:
Iris images captured in less-constrained environments often suffer from adverse noise, challenging many existing segmentation algorithms. In this paper, we propose an eff...Show MoreMetadata
Abstract:
Iris images captured in less-constrained environments often suffer from adverse noise, challenging many existing segmentation algorithms. In this paper, we propose an efficient Hybrid Transformer U-Net (HTU-Net) to address this dilemma. Unlike previous studies that only focus on utilizing popular CNN technology to predict iris masks accurately, HTU-Net can simultaneously obtain segmentation masks and parameterized pupillary and limbic boundaries by a multi-task network, further enabling CNN-based iris segmentation to be applied in any regular iris recognition systems. We explore the application of the Transformer in iris segmentation and propose a hybrid encoder that employs convolutional layers to extract local intensity features and the Transformer to capture long-range associative information. For decoding, we adopt a novel Multi-Head Dilated Attention to exploit the multi-scale contextual information by gating mechanism, thus emphasizing the important features and rendering powerful representations. Inspired by the consistent class characteristics of iris, we further devise a Pyramid Center-Aware Module to capture the global structural context of iris from a categorical perspective to improve performance. Experimental results show that our method, with fewer parameters than previous approaches, achieves competitive or new state-of-the-art performance in both iris segmentation and localization on three challenging iris datasets. Code will be released at https://github.com/Syloveslife/HTU-Net.
Date of Conference: 10-13 October 2022
Date Added to IEEE Xplore: 17 January 2023
ISBN Information: