I. Introduction
Medical image segmentation of anatomical structures and lesion regions is vital for the medical image analysis and clinical diagnosis. Convolutional neural networks (CNN) have achieved great progress on medical image segmentation [1], [2], [3], [4], [5]. However, these fully-supervised methods depend on a large amount of labeled data. For medical image segmentation, pixel-wise annotation requires domain knowledge, and obtaining massive amounts of labeled images is expensive and time-consuming. Three medical image segmentation tasks are shown in Fig.1, and labeling the lesion or vessel masks is very time-consuming, especially for thin vessels of fundus images. Therefore, training deep models in the low data regime is a practical and challenging task for medical image segmentation.
Examples of three medical image segmentation tasks. The first row shows gastrointestinal polyp image, fundus image and dermoscopy image, respectively. The second row shows the corresponding ground truth masks, respectively. It can be seen from the fundus image that some thin vessels are very small in size, and the surrounding tissue in the polyp image is very similar in texture, and the pixel-level annotation is time-consuming.