Please use this identifier to cite or link to this item:
https://elib.vku.udn.vn/handle/123456789/6174| Title: | Few-Shot Retinal Vessel Semantic Segmentation Under Threshold-Based Co-training Dual Network |
| Authors: | Le, Tran Quoc Khanh Pham, Ha Hieu Nguyen, Hoang Thien Nguyen, Quan Le, Minh Huu Nhat Dinh, Quang Vinh |
| Keywords: | Co-training Few-shot learning Pseudo-labeling Retinal vessel segmentation |
| Issue Date: | Jan-2026 |
| Publisher: | Springer Nature |
| Abstract: | Retinal vessel segmentation plays an important role in the early diagnosis and monitoring of many ocular and systemic diseases. However, labeled medical imaging data is scarce, costly, and requires pixel-level precision, making few-shot learning a promising solution to this challenge. This paper presents a novel threshold-based co-training framework using dual network for few-shot retinal vessel segmentation. Specifically, the method employs two segmentation models initialized with different parameters, which collaboratively learn by leveraging high-confidence pseudo-labels generated through a thresholding mechanism. The proposed method balances segmentation accuracy and robustness by integrating binary cross-entropy and Dice loss, effectively minimizing noise and uncertainty. Evaluation on the CHASE_DB1 dataset shows superior performance compared to state-of-the-art methods, achieving improvements in accuracy, sensitivity, specificity, and Dice score. These findings highlight the potential of threshold-based co-trained dual network for efficient and accurate retinal vessel segmentation using limited data. |
| Description: | Lecture Notes in Networks and Systems (LNNS,volume 1581); The 14th Conference on Information Technology and Its Applications (CITA 2025) ; pp: 831-842 |
| URI: | https://doi.org/10.1007/978-3-032-00972-2_61 https://elib.vku.udn.vn/handle/123456789/6174 |
| ISBN: | 978-3-032-00971-5 (p) 978-3-032-00972-2 (e) |
| Appears in Collections: | CITA 2025 (International) |
Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.