TY - JOUR
T1 - ZAP-2.5DSAM: zero additional parameters advancing 2.5D SAM adaptation to 3D tumor segmentation
AU - Guo, Cai
AU - Jin, Yuxi
AU - Tao, Bishenghui
AU - Li, Jianzhong
AU - Dai, Hong Ning
AU - Li, Ping
N1 - Publisher Copyright:
© The Author(s), under exclusive licence to Springer-Verlag GmbH Germany, part of Springer Nature 2025.
Funding Information:
This work was supported by the National Natural Science Foundation of China (12305409); the Guangdong Basic and Applied Basic Research Foundation (2021A1515011091, 2022A1515110696, 2024A1515011825, 2025A1515012945, 2025A1515012911); the Educational Commission of Guangdong Province (2020ZDZX3056); Research Matching Grant Scheme from the University Grants Council of the Hong Kong Special Administrative Region, China (No. 2024/3003) and Gekko Lab; the Shenzhen Science and Technology Program of China (JCYJ20230807145004008, JCYJ20240813155826035); and the Doctor Starting Fund of Hanshan Normal University, China (QD202324).
PY - 2025/10
Y1 - 2025/10
N2 - The segment anything model (SAM) demonstrated outstanding performance in 2D segmentation tasks, exhibiting robust generalization to natural images through its prompt-driven design. However, due to the lack of volumetric spatial information modeling and the domain gap between nature and medical images, its direct application to 3D medical image segmentation is suboptimal. Existing approaches to adapting SAM for 3D segmentation typically involve architectural adjustments by integrating additional components, thereby increasing trainable parameters and requiring higher GPU memory during fine-tuning. Moreover, retraining the prompt encoder may result in degraded spatial localization, especially when annotated data is scarce. To address these limitations, we propose ZAP-2.5DSAM, a parameter-efficient fine-tuning framework, which effectively extends the segmentation capacity of SAM to 3D medical images through a 2.5D decomposition scheme without introducing any additional adapter modules. Our method fine-tunes only 3.51M parameters from the original SAM, significantly reducing GPU memory requirements during training. Extensive experiments on multiple 3D tumor segmentation benchmarks demonstrate that ZAP-2.5DSAM achieves superior segmentation accuracy compared to conventional fine-tuning methods. Our code and models are available at: https://github.com/CaiGuoHS/ZAP-2.5DSAM.git.
AB - The segment anything model (SAM) demonstrated outstanding performance in 2D segmentation tasks, exhibiting robust generalization to natural images through its prompt-driven design. However, due to the lack of volumetric spatial information modeling and the domain gap between nature and medical images, its direct application to 3D medical image segmentation is suboptimal. Existing approaches to adapting SAM for 3D segmentation typically involve architectural adjustments by integrating additional components, thereby increasing trainable parameters and requiring higher GPU memory during fine-tuning. Moreover, retraining the prompt encoder may result in degraded spatial localization, especially when annotated data is scarce. To address these limitations, we propose ZAP-2.5DSAM, a parameter-efficient fine-tuning framework, which effectively extends the segmentation capacity of SAM to 3D medical images through a 2.5D decomposition scheme without introducing any additional adapter modules. Our method fine-tunes only 3.51M parameters from the original SAM, significantly reducing GPU memory requirements during training. Extensive experiments on multiple 3D tumor segmentation benchmarks demonstrate that ZAP-2.5DSAM achieves superior segmentation accuracy compared to conventional fine-tuning methods. Our code and models are available at: https://github.com/CaiGuoHS/ZAP-2.5DSAM.git.
KW - Segment anything model (SAM)
KW - Tumor segmentation
KW - Zero additional parameters (ZAP)
KW - 2.5D fine-tuning framework
UR - https://www.scopus.com/pages/publications/105010578924
U2 - 10.1007/s00371-025-04092-4
DO - 10.1007/s00371-025-04092-4
M3 - Journal article
AN - SCOPUS:105010578924
SN - 0178-2789
VL - 41
SP - 11139
EP - 11149
JO - Visual Computer
JF - Visual Computer
IS - 13
ER -