Enhancing Few-Shot Medical Image Segmentation with Refined Prototypes

Show simple item record

dc.contributor.author Ishrak, Isaba
dc.contributor.author Nushra, Sabah
dc.contributor.author Anika, Fairoz
dc.date.accessioned 2025-05-30T10:29:02Z
dc.date.available 2025-05-30T10:29:02Z
dc.date.issued 2024-06-30
dc.identifier.citation [1] F. Aurenhammer, “Voronoi diagrams—a survey of a fundamental geometric data structure,” ACM Computing Surveys (CSUR), vol. 23, no. 3, pp. 345–405, 1991. [2] H. Cao, Y. Wang, J. Chen, et al., “Swin-unet: Unet-like pure transformer for medical image segmentation,” in Proceedings of the Workshop of European Con ference on Computer Vision, 2023, pp. 205–. [3] J. Chen, Y. Lu, Q. Yu, et al., “Transunet: Transformers make strong encoders for medical image segmentation,” arXiv preprint arXiv:2102.04306, 2021. [4] O. Cïcek, A. Abdulkadir, S. S. Lienkamp, T. Brox, and O. Ronneberger, “3d u net: Learning dense volumetric segmentation from sparse annotation,” in In ternational Conference on Medical Image Computing and Computer-Assisted In tervention, Springer, 2016, pp. 424–432. [5] N. Dong and E. P. Xing, “Few-shot semantic segmentation with prototype learn ing,” in BMVC, vol. 2, 2018. [6] A. Dosovitskiy, L. Beyer, A. Kolesnikov, D. Weissenborn, et al., “An image is worth 16x16 words: Transformers for image recognition at scale,” in Interna tional Conference on Learning Representations (ICLR), 2021. [7] C. Finn, P. Abbeel, and S. Levine, “Model-agnostic meta-learning for fast adap tation of deep networks,” in Proceedings of the 34th International Conference on Machine Learning (ICML), 2017, pp. 1126–1135. [8] H. Gharoun, F. Momenifar, F. Chen, and A. H. Gandomi, “Meta-learning ap proaches for few-shot learning: A survey of recent advances,” arXiv preprint arXiv:2303.07502, 2023. [9] S. Hansen, S. Gautam, R. Jenssen, and M. Kampffmeyer, “Anomaly detection inspired few-shot medical image segmentation through self-supervision with supervoxels,” Medical Image Analysis, vol. 78, p. 102 385, 2022. 43 [10] K. He, X. Zhang, S. Ren, and J. Sun, “Deep residual learning for image recog nition,” in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, 2016, pp. 770–778. [11] B. Hui, P. Zhu, Q. Hu, and Q. Wang, “Self-attention relation network for few shot learning,” in 2019 IEEE international conference on Multimedia & Expo Workshops (ICMEW), IEEE, 2019, pp. 198–203. [12] F. Isensee, J. Petersen, A. Klein, et al., “Nnu-net: Self-adapting framework for u net-based medical image segmentation,” arXiv preprint arXiv:1809.10486, 2018. [13] A. E. Kavur, N. S. Gezer, M. Barıs¸, et al., “Chaos challenge-combined (ct-mr) healthy abdominal organ segmentation,” Medical Image Analysis, vol. 69, p. 101 950, 2021. [14] B. Landman, Z. Xu, J. Igelsias, M. Styner, T. Langerak, and A. Klein, “Miccai multi-atlas labeling beyond the cranial vault–workshop and challenge,” in MIC CAI—Workshop Challenge, 2015, p. 12. [15] X. Li, L. Yu, H. Chen, et al., “Transformation-consistent self-ensembling model for semi-supervised medical image segmentation,” IEEE Transactions on Neu ral Networks and Learning Systems, vol. 32, no. 2, pp. 523–534, 2020. doi: 10. 1109/TNNLS.2020.2975191. [16] A. Lin, B. Chen, J. Xu, Z. Zhang, G. Lu, and D. Zhang, “Ds-transunet: Dual swin transformer u-net for medical image segmentation,” IEEE Transactions on Instrumentation and Measurement, vol. 71, pp. 1–15, 2022. [17] Y. Lin, Y. Chen, K.-T. Cheng, and H. Chen, “Few shot medical image segmen tation with cross attention transformer,” arXiv preprint arXiv:2303.13867, 2023. [18] J. Liu and Y. Qin, “Prototype refinement network for few-shot segmentation,” arXiv preprint arXiv:2002.03579, 2020. [19] Y. Liu, X. Zhang, S. Zhang, and X. He, “Part-aware prototype network for few shot semantic segmentation,” in European Conference on Computer Vision, Springer, 2020, pp. 142–158. [20] J. Long, E. Shelhamer, and T. Darrell, “Fully convolutional networks for seman tic segmentation,” in Proceedings of the IEEE conference on computer vision and pattern recognition, 2015, pp. 3431–3440. [21] Z. Lu, S. He, X. Zhu, L. Zhang, Y.-Z. Song, and T. Xiang, “Simpler is better: Few shot semantic segmentation with classifier weight transformer,” in Proceedings of the IEEE/CVF International Conference on Computer Vision, 2021, pp. 8741–. 44 [22] S. Matsumi and K. Yamada, “Few-shot learning based on metric learning using class augmentation,” in 2020 25th International Conference on Pattern Recogni tion (ICPR), 2021, pp. 196–201. doi: 10.1109/ICPR48806.2021.9411993. [23] S. Mehta, E. Mercan, J. Bartlett, D. Weaver, J. G. Elmore, and L. Shapiro, “Y-net: Joint segmentation and classification for diagnosis of breast biopsy images,” in International Conference on Medical Image Computing and Computer-Assisted Intervention, Springer, 2018, pp. 893–901. [24] F. Milletari, N. Navab, and S.-A. Ahmadi, “V-net: Fully convolutional neural networks for volumetric medical image segmentation,” in 2016 Fourth Interna tional Conference on 3D Vision (3DV), IEEE, 2016, pp. 565–571. [25] O. Oktay, J. Schlemper, L. L. Folgoc, et al., “Attention u-net: Learning where to look for the pancreas,” in Medical Imaging with Deep Learning (MIDL), 2018. [26] O. Oktay, J. Schlemper, L. Le Folgoc, et al., “Attention u-net: Learning where to look for the pancreas,” arXiv preprint arXiv:1804.03999, 2018. [27] C. Ouyang, C. Biffi, C. Chen, T. Kart, H. Qiu, and D. Rueckert, “Self-supervision with superpixels: Training few-shot medical image segmentation without an notation,” in Computer Vision–ECCV 2020: 16th European Conference, Glasgow, UK, August 23–28, 2020, Proceedings, Part XXIX 16, Springer, 2020, pp. 762–780. [28] C. Ouyang, C. Biffi, C. Chen, T. Kart, H. Qiu, and D. Rueckert, “Self-supervised learning for few-shot medical image segmentation,” IEEE Transactions on Med ical Imaging, vol. 41, no. 7, pp. 1837–1848, 2022. [29] X. Ouyang, J. Liu, and L. Zhang, “Ssl-alpnet: Self-supervised few-shot learn ing for medical image segmentation,” IEEE Transactions on Medical Imaging, vol. 39, no. 10, pp. 3450–3460, 2020. doi: 10.1109/TMI.2020.3015105. [30] O. Ronneberger, P. Fischer, and T. Brox, “U-net: Convolutional networks for biomedical image segmentation,” in Medical Image Computing and Computer Assisted Intervention–MICCAI 2015: 18th International Conference, Munich, Ger many, October 5-9, 2015, Proceedings, Part III 18, Springer, 2015, pp. 234–241. [31] A. Roy, S. Siddiqui, S. Pölsterl, N. Navab, and C. Wachinger, “’squeeze excite’ guided few-shot segmentation of volumetric images,” Medical Image Analysis, vol. 59, p. 101 587, 2020. [32] Q. Shen, Y. Li, J. Jin, and B. Liu, “Q-net: Query-informed few-shot medical im age segmentation,” arXiv preprint arXiv:2208.11451, 2022. [33] J. Snell, K. Swersky, and R. Zemel, “Prototypical networks for few-shot learn ing,” in Advances in Neural Information Processing Systems, vol. 30, 2017. 45 [34] L. Sun, C. Li, X. Ding, et al., “Few-shot medical image segmentation using a global correlation network with discriminative embedding,” Computers in Bi ology and Medicine, vol. 140, p. 105 067, 2022. [35] K. Wang, J. H. Liew, Y. Zou, D. Zhou, and J. Feng, “Panet: Few-shot image se mantic segmentation with prototype alignment,” in Proc. of the IEEE/CVF In ternational Conference on Computer Vision, 2019, pp. 9197–9206. [36] R. Wang, Q. Zhou, and G. Zheng, “Few-shot medical image segmentation reg ularized with self-reference and contrastive learning,” in Proceedings of the In ternational Conference on Medical Image Computing and Computer Assisted In tervention, 2022, pp. 514–523. [37] W. Wang, C. Chen, M. Ding, H. Yu, S. Zha, and J. Li, “Transbts: Multimodal brain tumor segmentation using transformer,” in Proceedings of the Interna tional Conference on Medical Image Computing and Computer-Assisted Inter vention, 2021, pp. 109–119. [38] H. Wu, F. Xiao, and C. Liang, “Dual contrastive learning with anatomical aux iliary supervision for few-shot medical image segmentation,” in European Con ference on Computer Vision, 2022, pp. 417–434. [39] Z. Yu, L. Chen, Z. Cheng, and J. Luo, “Transmatch: A transfer-learning scheme for semi-supervised few-shot learning,” in Proceedings of the IEEE/CVF confer ence on computer vision and pattern recognition, 2020, pp. 12 856–12 864. [40] Y. Zhang, H. Liu, and Q. Hu, “Transfuse: Fusing transformers and cnns for medical image segmentation,” in Proceedings of the International Conference on Medical Image Computing and Computer-Assisted Intervention, 2021, pp. 14–24. [41] Z. Zhang, Y. Wei, Y. Yang, Y. Yang, et al., “Sg-one: Similarity guidance network for one-shot semantic segmentation,” IEEE Transactions on Cybernetics, vol. 50, no. 9, pp. 3855–3865, 2020. doi: 10.1109/TCYB.2020.2979134. [42] N. Zhao, S. Qiao, and J. Peng, “Few-shot learning with uncertainty-aware at tention,” in Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV), 2021, pp. 4911–4920. [43] H. Zhou, Y. Sun, and Q. Dong, “Region-enhanced prototypical transformer for few-shot medical image segmentation,” in International Conference on Medical Image Computing and Computer-Assisted Intervention, Springer, 2021, pp. 239– 249. doi: 10.1007/978-3-030-87234-2_23. [44] Z. Zhou, M. M. R. Siddiquee, N. Tajbakhsh, and J. Liang, “Unet++: Redesigning skip connections to exploit multiscale features in image segmentation,” IEEE transactions on medical imaging, vol. 39, no. 6, pp. 1856–1867, 2019. 46 [45] Y. Zhu, S. Wang, T. Xin, and H. Zhang, “Few-shot medical image segmentation via a region-enhanced prototypical transformer,” in International Conference on Medical Image Computing and Computer-Assisted Intervention, Springer, 2023, pp. 271–280. [46] Y. Zhu, S. Wang, T. Xin, Z. Zhang, and H. Zhang, “Partition-a-medical-image: Extracting multiple representative sub-regions for few-shot medical image seg mentation,” arXiv preprint arXiv:2309.11172, 2023. [47] X. Zhuang, “Multivariate mixture model for myocardial segmentation combin ing multi-source images,” IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 41, no. 12, pp. 2933–2946, 2018. en_US
dc.identifier.uri http://hdl.handle.net/123456789/2409
dc.description Supervised by Prof. Dr. Hasanul Kabir, Co-Supervised By Mr. Sabbir Ahmed, Assistant Professor, Department of Computer Science and Engineering (CSE) Islamic University of Technology (IUT) Board Bazar, Gazipur, Bangladesh This thesis is submitted in partial fulfillment of the requirement for the degree of Bachelor of Science in Computer Science and Engineering, 2024 en_US
dc.description.abstract Deep learning has achieved tremendous success in computer vision, while medical image segmentation (MIS) remains a challenge, due to the scarcity of data anno tations. Deep learning-based medical image segmentation methods are recognized as data-hungry techniques that require large amounts of data with manual annota tions. However, manual annotation is expensive in the field of medical image analysis, which requires domain-specific expertise. To address this challenge, few-shot learn ing has the potential to learn new classes from only a few labeled examples. In this work, we propose a framework for few-shot medical image segmentation based on cross-masked attention Transformer and Region-enhanced Prototypical Transformer. Our proposed network mines the correlations between the support image and query image, limiting them to focus only on useful foreground information and boosting the representation capacity of both the support prototype and query features. To mit igate the effects of large intra-class diversity, we further design a subdivision strategy to produce a collection of regional prototypes from the foreground of the support, and a self-selection mechanism is proposed to incorporate into the Bias-alleviated Trans former (BaT) block to suppress or remove interferences present in the query prototype and regional support prototypes. Thus, we have combined the feature enhancement and prototype generation in the Enhanced Prototype Generation EPG module which will iteratively update the generated query mask by taking the query mask gener ated in the previous iteration and finally produce more accurate global prototypes for Few-Shot Medical Image Segmentation. We conducted experiments on three publicly available medical image datasets, Abd-CT, Abd-MRI, and CMR to show the segmen tation results of the available state-of-the-art methods. Our experiments yielded final mean Dice scores of 50.82% on Abd-CT, 45.82% on Abd-MRI, and 75.08% on the CMR dataset, demonstrating competitive performance across these challenging datasets. en_US
dc.language.iso en en_US
dc.publisher Department of Computer Science and Engineering(CSE), Islamic University of Technology(IUT), Board Bazar, Gazipur-1704, Bangladesh en_US
dc.subject 1. Fewshot 2. Medical image segmentation 3. Prototypical network 4. Abdominal MRI 5. FSMS en_US
dc.title Enhancing Few-Shot Medical Image Segmentation with Refined Prototypes en_US
dc.type Thesis en_US


Files in this item

This item appears in the following Collection(s)

Show simple item record

Search IUT Repository


Advanced Search

Browse

My Account

Statistics