TY - JOUR
T1 - Volumetric medical image segmentation via fully 3D adaptation of Segment Anything Model
AU - Lin, Haoneng
AU - Zou, Jing
AU - Deng, Sen
AU - Wong, Ka Po
AU - Aviles-Rivero, Angelica I.
AU - Fan, Yiting
AU - Lee, Alex Pui Wai
AU - Hu, Xiaowei
AU - Qin, Jing
N1 - Publisher Copyright:
© 2024 Nalecz Institute of Biocybernetics and Biomedical Engineering of the Polish Academy of Sciences
PY - 2025/1/1
Y1 - 2025/1/1
N2 - The Segment Anything Model (SAM) exhibits exceptional generalization capabilities in diverse domains, owing to its interactive learning mechanism designed for precise image segmentation. However, applying SAM to out-of-distribution domains, especially in 3D medical image segmentation, poses challenges. Existing methods for adapting 2D segmentation models to 3D medical data treat 3D volumes as a mere stack of 2D slices. The essential inter-slice information, which is pivotal to faithful 3D medical image segmentation tasks, is unfortunately neglected. In this work, we present the 3D Medical SAM-Adapter (3DMedSAM), a pioneer cross-dimensional adaptation, leveraging SAM's pre-trained knowledge while accommodating the unique characteristics of 3D medical data. Firstly, to bridge the dimensional gap from 2D to 3D, we design a novel module to replace SAM's patch embedding, ensuring a seamless transition into 3D image processing and recognition. Besides, we incorporate a 3D Adapter while maintaining the majority of pre-training parameters frozen, enriching deep features with abundant 3D spatial information and achieving efficient fine-tuning. Given the diverse scales of anomalies present in medical images, we also devised a multi-scale 3D mask decoder to elevate the network's proficiency in medical image segmentation. Through various experiments, we showcase the effectiveness of 3DMedSAM in achieving accurate and robust 3D segmentation on both single-target segmentation and multi-organ segmentation tasks, surpassing the limitations of current methods.
AB - The Segment Anything Model (SAM) exhibits exceptional generalization capabilities in diverse domains, owing to its interactive learning mechanism designed for precise image segmentation. However, applying SAM to out-of-distribution domains, especially in 3D medical image segmentation, poses challenges. Existing methods for adapting 2D segmentation models to 3D medical data treat 3D volumes as a mere stack of 2D slices. The essential inter-slice information, which is pivotal to faithful 3D medical image segmentation tasks, is unfortunately neglected. In this work, we present the 3D Medical SAM-Adapter (3DMedSAM), a pioneer cross-dimensional adaptation, leveraging SAM's pre-trained knowledge while accommodating the unique characteristics of 3D medical data. Firstly, to bridge the dimensional gap from 2D to 3D, we design a novel module to replace SAM's patch embedding, ensuring a seamless transition into 3D image processing and recognition. Besides, we incorporate a 3D Adapter while maintaining the majority of pre-training parameters frozen, enriching deep features with abundant 3D spatial information and achieving efficient fine-tuning. Given the diverse scales of anomalies present in medical images, we also devised a multi-scale 3D mask decoder to elevate the network's proficiency in medical image segmentation. Through various experiments, we showcase the effectiveness of 3DMedSAM in achieving accurate and robust 3D segmentation on both single-target segmentation and multi-organ segmentation tasks, surpassing the limitations of current methods.
KW - 3D medical image segmentation
KW - Cross-dimensional adaptation
KW - Segment Anything Model
UR - https://www.scopus.com/pages/publications/85211021964
U2 - 10.1016/j.bbe.2024.11.001
DO - 10.1016/j.bbe.2024.11.001
M3 - Journal article
AN - SCOPUS:85211021964
SN - 0208-5216
VL - 45
SP - 1
EP - 10
JO - Biocybernetics and Biomedical Engineering
JF - Biocybernetics and Biomedical Engineering
IS - 1
ER -