dc.identifier.citation |
[1] F. Aurenhammer, “Voronoi diagrams—a survey of a fundamental geometric data structure,” ACM Computing Surveys (CSUR), vol. 23, no. 3, pp. 345–405, 1991. [2] H. Cao, Y. Wang, J. Chen, et al., “Swin-unet: Unet-like pure transformer for medical image segmentation,” in Proceedings of the Workshop of European Con ference on Computer Vision, 2023, pp. 205–. [3] J. Chen, Y. Lu, Q. Yu, et al., “Transunet: Transformers make strong encoders for medical image segmentation,” arXiv preprint arXiv:2102.04306, 2021. [4] O. Cïcek, A. Abdulkadir, S. S. Lienkamp, T. Brox, and O. Ronneberger, “3d u net: Learning dense volumetric segmentation from sparse annotation,” in In ternational Conference on Medical Image Computing and Computer-Assisted In tervention, Springer, 2016, pp. 424–432. [5] N. Dong and E. P. Xing, “Few-shot semantic segmentation with prototype learn ing,” in BMVC, vol. 2, 2018. [6] A. Dosovitskiy, L. Beyer, A. Kolesnikov, D. Weissenborn, et al., “An image is worth 16x16 words: Transformers for image recognition at scale,” in Interna tional Conference on Learning Representations (ICLR), 2021. [7] C. Finn, P. Abbeel, and S. Levine, “Model-agnostic meta-learning for fast adap tation of deep networks,” in Proceedings of the 34th International Conference on Machine Learning (ICML), 2017, pp. 1126–1135. [8] H. Gharoun, F. Momenifar, F. Chen, and A. H. Gandomi, “Meta-learning ap proaches for few-shot learning: A survey of recent advances,” arXiv preprint arXiv:2303.07502, 2023. [9] S. Hansen, S. Gautam, R. Jenssen, and M. Kampffmeyer, “Anomaly detection inspired few-shot medical image segmentation through self-supervision with supervoxels,” Medical Image Analysis, vol. 78, p. 102 385, 2022. 43 [10] K. He, X. Zhang, S. Ren, and J. Sun, “Deep residual learning for image recog nition,” in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, 2016, pp. 770–778. [11] B. Hui, P. Zhu, Q. Hu, and Q. Wang, “Self-attention relation network for few shot learning,” in 2019 IEEE international conference on Multimedia & Expo Workshops (ICMEW), IEEE, 2019, pp. 198–203. [12] F. Isensee, J. Petersen, A. Klein, et al., “Nnu-net: Self-adapting framework for u net-based medical image segmentation,” arXiv preprint arXiv:1809.10486, 2018. [13] A. E. Kavur, N. S. Gezer, M. Barıs¸, et al., “Chaos challenge-combined (ct-mr) healthy abdominal organ segmentation,” Medical Image Analysis, vol. 69, p. 101 950, 2021. [14] B. Landman, Z. Xu, J. Igelsias, M. Styner, T. Langerak, and A. Klein, “Miccai multi-atlas labeling beyond the cranial vault–workshop and challenge,” in MIC CAI—Workshop Challenge, 2015, p. 12. [15] X. Li, L. Yu, H. Chen, et al., “Transformation-consistent self-ensembling model for semi-supervised medical image segmentation,” IEEE Transactions on Neu ral Networks and Learning Systems, vol. 32, no. 2, pp. 523–534, 2020. doi: 10. 1109/TNNLS.2020.2975191. [16] A. Lin, B. Chen, J. Xu, Z. Zhang, G. Lu, and D. Zhang, “Ds-transunet: Dual swin transformer u-net for medical image segmentation,” IEEE Transactions on Instrumentation and Measurement, vol. 71, pp. 1–15, 2022. [17] Y. Lin, Y. Chen, K.-T. Cheng, and H. Chen, “Few shot medical image segmen tation with cross attention transformer,” arXiv preprint arXiv:2303.13867, 2023. [18] J. Liu and Y. Qin, “Prototype refinement network for few-shot segmentation,” arXiv preprint arXiv:2002.03579, 2020. [19] Y. Liu, X. Zhang, S. Zhang, and X. He, “Part-aware prototype network for few shot semantic segmentation,” in European Conference on Computer Vision, Springer, 2020, pp. 142–158. [20] J. Long, E. Shelhamer, and T. Darrell, “Fully convolutional networks for seman tic segmentation,” in Proceedings of the IEEE conference on computer vision and pattern recognition, 2015, pp. 3431–3440. [21] Z. Lu, S. He, X. Zhu, L. Zhang, Y.-Z. Song, and T. Xiang, “Simpler is better: Few shot semantic segmentation with classifier weight transformer,” in Proceedings of the IEEE/CVF International Conference on Computer Vision, 2021, pp. 8741–. 44 [22] S. Matsumi and K. Yamada, “Few-shot learning based on metric learning using class augmentation,” in 2020 25th International Conference on Pattern Recogni tion (ICPR), 2021, pp. 196–201. doi: 10.1109/ICPR48806.2021.9411993. [23] S. Mehta, E. Mercan, J. Bartlett, D. Weaver, J. G. Elmore, and L. Shapiro, “Y-net: Joint segmentation and classification for diagnosis of breast biopsy images,” in International Conference on Medical Image Computing and Computer-Assisted Intervention, Springer, 2018, pp. 893–901. [24] F. Milletari, N. Navab, and S.-A. Ahmadi, “V-net: Fully convolutional neural networks for volumetric medical image segmentation,” in 2016 Fourth Interna tional Conference on 3D Vision (3DV), IEEE, 2016, pp. 565–571. [25] O. Oktay, J. Schlemper, L. L. Folgoc, et al., “Attention u-net: Learning where to look for the pancreas,” in Medical Imaging with Deep Learning (MIDL), 2018. [26] O. Oktay, J. Schlemper, L. Le Folgoc, et al., “Attention u-net: Learning where to look for the pancreas,” arXiv preprint arXiv:1804.03999, 2018. [27] C. Ouyang, C. Biffi, C. Chen, T. Kart, H. Qiu, and D. Rueckert, “Self-supervision with superpixels: Training few-shot medical image segmentation without an notation,” in Computer Vision–ECCV 2020: 16th European Conference, Glasgow, UK, August 23–28, 2020, Proceedings, Part XXIX 16, Springer, 2020, pp. 762–780. [28] C. Ouyang, C. Biffi, C. Chen, T. Kart, H. Qiu, and D. Rueckert, “Self-supervised learning for few-shot medical image segmentation,” IEEE Transactions on Med ical Imaging, vol. 41, no. 7, pp. 1837–1848, 2022. [29] X. Ouyang, J. Liu, and L. Zhang, “Ssl-alpnet: Self-supervised few-shot learn ing for medical image segmentation,” IEEE Transactions on Medical Imaging, vol. 39, no. 10, pp. 3450–3460, 2020. doi: 10.1109/TMI.2020.3015105. [30] O. Ronneberger, P. Fischer, and T. Brox, “U-net: Convolutional networks for biomedical image segmentation,” in Medical Image Computing and Computer Assisted Intervention–MICCAI 2015: 18th International Conference, Munich, Ger many, October 5-9, 2015, Proceedings, Part III 18, Springer, 2015, pp. 234–241. [31] A. Roy, S. Siddiqui, S. Pölsterl, N. Navab, and C. Wachinger, “’squeeze excite’ guided few-shot segmentation of volumetric images,” Medical Image Analysis, vol. 59, p. 101 587, 2020. [32] Q. Shen, Y. Li, J. Jin, and B. Liu, “Q-net: Query-informed few-shot medical im age segmentation,” arXiv preprint arXiv:2208.11451, 2022. [33] J. Snell, K. Swersky, and R. Zemel, “Prototypical networks for few-shot learn ing,” in Advances in Neural Information Processing Systems, vol. 30, 2017. 45 [34] L. Sun, C. Li, X. Ding, et al., “Few-shot medical image segmentation using a global correlation network with discriminative embedding,” Computers in Bi ology and Medicine, vol. 140, p. 105 067, 2022. [35] K. Wang, J. H. Liew, Y. Zou, D. Zhou, and J. Feng, “Panet: Few-shot image se mantic segmentation with prototype alignment,” in Proc. of the IEEE/CVF In ternational Conference on Computer Vision, 2019, pp. 9197–9206. [36] R. Wang, Q. Zhou, and G. Zheng, “Few-shot medical image segmentation reg ularized with self-reference and contrastive learning,” in Proceedings of the In ternational Conference on Medical Image Computing and Computer Assisted In tervention, 2022, pp. 514–523. [37] W. Wang, C. Chen, M. Ding, H. Yu, S. Zha, and J. Li, “Transbts: Multimodal brain tumor segmentation using transformer,” in Proceedings of the Interna tional Conference on Medical Image Computing and Computer-Assisted Inter vention, 2021, pp. 109–119. [38] H. Wu, F. Xiao, and C. Liang, “Dual contrastive learning with anatomical aux iliary supervision for few-shot medical image segmentation,” in European Con ference on Computer Vision, 2022, pp. 417–434. [39] Z. Yu, L. Chen, Z. Cheng, and J. Luo, “Transmatch: A transfer-learning scheme for semi-supervised few-shot learning,” in Proceedings of the IEEE/CVF confer ence on computer vision and pattern recognition, 2020, pp. 12 856–12 864. [40] Y. Zhang, H. Liu, and Q. Hu, “Transfuse: Fusing transformers and cnns for medical image segmentation,” in Proceedings of the International Conference on Medical Image Computing and Computer-Assisted Intervention, 2021, pp. 14–24. [41] Z. Zhang, Y. Wei, Y. Yang, Y. Yang, et al., “Sg-one: Similarity guidance network for one-shot semantic segmentation,” IEEE Transactions on Cybernetics, vol. 50, no. 9, pp. 3855–3865, 2020. doi: 10.1109/TCYB.2020.2979134. [42] N. Zhao, S. Qiao, and J. Peng, “Few-shot learning with uncertainty-aware at tention,” in Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV), 2021, pp. 4911–4920. [43] H. Zhou, Y. Sun, and Q. Dong, “Region-enhanced prototypical transformer for few-shot medical image segmentation,” in International Conference on Medical Image Computing and Computer-Assisted Intervention, Springer, 2021, pp. 239– 249. doi: 10.1007/978-3-030-87234-2_23. [44] Z. Zhou, M. M. R. Siddiquee, N. Tajbakhsh, and J. Liang, “Unet++: Redesigning skip connections to exploit multiscale features in image segmentation,” IEEE transactions on medical imaging, vol. 39, no. 6, pp. 1856–1867, 2019. 46 [45] Y. Zhu, S. Wang, T. Xin, and H. Zhang, “Few-shot medical image segmentation via a region-enhanced prototypical transformer,” in International Conference on Medical Image Computing and Computer-Assisted Intervention, Springer, 2023, pp. 271–280. [46] Y. Zhu, S. Wang, T. Xin, Z. Zhang, and H. Zhang, “Partition-a-medical-image: Extracting multiple representative sub-regions for few-shot medical image seg mentation,” arXiv preprint arXiv:2309.11172, 2023. [47] X. Zhuang, “Multivariate mixture model for myocardial segmentation combin ing multi-source images,” IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 41, no. 12, pp. 2933–2946, 2018. |
en_US |