Abstract
Transfer learning has witnessed a recent surge of interest after proving successful in multiple applications. However, it highly relies on the quantity of annotated data. Constrained by the labor cost and expertise, it is hard to annotate sufficient organs and tumors at the voxel level for medical image segmentation. Consequently, most bench-mark datasets were collected for the segmentation of only one type of organ and/or tumor, and all task-irrelevant organs and tumors were annotated as the background. We aim to make use of these partially but plentifully labeled datasets to boost the segmentation performance of the annotation-limited KiTS21 segmentation task. To this end, we first construct a general medical image segmentation model that learns to segment these partially labeled organs or tumors. Then we transfer its pre-trained weights to a specific downstream task, i.e., KiTS21. The primary experiments demonstrate the effectiveness of the proposed transfer learning strategy. Our method achieves 0.890 Dice score, 0.805 SurfaceDice, and 0.822 Tumor Dice in the KiTS21 challenge.
X. Yang and J. Zhang—Equal contribution. This work was done during Xi Yang’s internship at JD Explore Academy.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
References
Bilic, P., et al.: The liver tumor segmentation benchmark (lits). arXiv preprint arXiv:1901.04056 (2019)
Chen, L.C., Papandreou, G., Kokkinos, I., Murphy, K., Yuille, A.L.: Deeplab: semantic image segmentation with deep convolutional nets, atrous convolution, and fully connected CRFs. IEEE Trans. Pattern Anal. Mach. Intell. 40(4), 834–848 (2017)
He, K., Zhang, X., Ren, S., Sun, J.: Deep residual learning for image recognition. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp. 770–778 (2016)
Heller, N., et al.: The kits19 challenge data: 300 kidney tumor cases with clinical context, CT semantic segmentations, and surgical outcomes. arXiv preprint arXiv:1904.00445 (2019)
Isensee, F., et al.: nnU-Net: self-adapting framework for U-Net-based medical image segmentation. arXiv preprint arXiv:1809.10486 (2018)
Simpson, A.L., et al.: A large annotated medical image dataset for the development and evaluation of segmentation algorithms. arXiv preprint arXiv:1902.09063 (2019)
Zhang, J., Xie, Y., Xia, Y., Shen, C.: DoDNet: learning to segment multi-organ and tumors from multiple partially labeled datasets. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 1195–1204 (2021)
Acknowledgment
This work was supported in part by the National Natural Science Foundation of China under Grants 61771397 and 62171377, and in part by the Science and Technology Innovation Committee of Shenzhen Municipality, China, under Grants JCYJ20180306171334997.
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2022 Springer Nature Switzerland AG
About this paper
Cite this paper
Yang, X., Zhang, J., Zhang, J., Xia, Y. (2022). Transfer Learning for KiTS21 Challenge. In: Heller, N., Isensee, F., Trofimova, D., Tejpaul, R., Papanikolopoulos, N., Weight, C. (eds) Kidney and Kidney Tumor Segmentation. KiTS 2021. Lecture Notes in Computer Science, vol 13168. Springer, Cham. https://doi.org/10.1007/978-3-030-98385-7_21
Download citation
DOI: https://doi.org/10.1007/978-3-030-98385-7_21
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-030-98384-0
Online ISBN: 978-3-030-98385-7
eBook Packages: Computer ScienceComputer Science (R0)