<?xml version="1.0" encoding="UTF-8"?><!DOCTYPE article PUBLIC "-//NLM//DTD Journal Publishing DTD v2.0 20040830//EN" "journalpublishing.dtd"><article xmlns:mml="http://www.w3.org/1998/Math/MathML" xmlns:xlink="http://www.w3.org/1999/xlink" dtd-version="2.0" xml:lang="en" article-type="research-article"><front><journal-meta><journal-id journal-id-type="nlm-ta">JMIR AI</journal-id><journal-id journal-id-type="publisher-id">ai</journal-id><journal-id journal-id-type="index">41</journal-id><journal-title>JMIR AI</journal-title><abbrev-journal-title>JMIR AI</abbrev-journal-title><issn pub-type="epub">2817-1705</issn><publisher><publisher-name>JMIR Publications</publisher-name><publisher-loc>Toronto, Canada</publisher-loc></publisher></journal-meta><article-meta><article-id pub-id-type="publisher-id">v4i1e72109</article-id><article-id pub-id-type="doi">10.2196/72109</article-id><article-categories><subj-group subj-group-type="heading"><subject>Original Paper</subject></subj-group></article-categories><title-group><article-title>Using Segment Anything Model 2 for Zero-Shot 3D Segmentation of Abdominal Organs in Computed Tomography Scans to Adapt Video Tracking Capabilities for 3D Medical Imaging: Algorithm Development and Validation</article-title></title-group><contrib-group><contrib contrib-type="author" corresp="yes"><name name-style="western"><surname>Yamagishi</surname><given-names>Yosuke</given-names></name><degrees>MD, MSc</degrees><xref ref-type="aff" rid="aff1">1</xref></contrib><contrib contrib-type="author"><name name-style="western"><surname>Hanaoka</surname><given-names>Shouhei</given-names></name><degrees>MD, PhD</degrees><xref ref-type="aff" rid="aff1">1</xref><xref ref-type="aff" rid="aff2">2</xref></contrib><contrib contrib-type="author"><name name-style="western"><surname>Kikuchi</surname><given-names>Tomohiro</given-names></name><degrees>MD, MPH, PhD</degrees><xref ref-type="aff" rid="aff3">3</xref><xref ref-type="aff" rid="aff4">4</xref></contrib><contrib contrib-type="author"><name name-style="western"><surname>Nakao</surname><given-names>Takahiro</given-names></name><degrees>MD, PhD</degrees><xref ref-type="aff" rid="aff4">4</xref></contrib><contrib contrib-type="author"><name name-style="western"><surname>Nakamura</surname><given-names>Yuta</given-names></name><degrees>MD, PhD</degrees><xref ref-type="aff" rid="aff4">4</xref></contrib><contrib contrib-type="author"><name name-style="western"><surname>Nomura</surname><given-names>Yukihiro</given-names></name><degrees>PhD</degrees><xref ref-type="aff" rid="aff4">4</xref><xref ref-type="aff" rid="aff5">5</xref></contrib><contrib contrib-type="author"><name name-style="western"><surname>Miki</surname><given-names>Soichiro</given-names></name><degrees>MD, PhD</degrees><xref ref-type="aff" rid="aff4">4</xref></contrib><contrib contrib-type="author"><name name-style="western"><surname>Yoshikawa</surname><given-names>Takeharu</given-names></name><degrees>MD, PhD</degrees><xref ref-type="aff" rid="aff4">4</xref></contrib><contrib contrib-type="author"><name name-style="western"><surname>Abe</surname><given-names>Osamu</given-names></name><degrees>MD, PhD</degrees><xref ref-type="aff" rid="aff1">1</xref><xref ref-type="aff" rid="aff2">2</xref></contrib></contrib-group><aff id="aff1"><institution>Division of Radiology and Biomedical Engineering, Graduate School of Medicine, The University of Tokyo</institution><addr-line>7-3-1 Hongo, Bunkyo-ku</addr-line><addr-line>Tokyo</addr-line><country>Japan</country></aff><aff id="aff2"><institution>Department of Radiology, The University of Tokyo Hospital</institution><addr-line>Tokyo</addr-line><country>Japan</country></aff><aff id="aff3"><institution>Department of Radiology, School of Medicine, Jichi Medical University</institution><addr-line>Shimotsuke</addr-line><country>Japan</country></aff><aff id="aff4"><institution>Department of Computational Diagnostic Radiology and Preventive Medicine, The University of Tokyo Hospital</institution><addr-line>Tokyo</addr-line><country>Japan</country></aff><aff id="aff5"><institution>Center for Frontier Medical Engineering, Chiba University</institution><addr-line>Chiba</addr-line><country>Japan</country></aff><contrib-group><contrib contrib-type="editor"><name name-style="western"><surname>Huo</surname><given-names>Yuankai</given-names></name></contrib></contrib-group><contrib-group><contrib contrib-type="reviewer"><name name-style="western"><surname>Zhang</surname><given-names>Yichi</given-names></name></contrib></contrib-group><author-notes><corresp>Correspondence to Yosuke Yamagishi, MD, MSc, Division of Radiology and Biomedical Engineering, Graduate School of Medicine, The University of Tokyo, 7-3-1 Hongo, Bunkyo-ku, Tokyo, 113-8655, Japan, 81 3-3815-5411; <email>yamagishi-yosuke0115@g.ecc.u-tokyo.ac.jp</email></corresp></author-notes><pub-date pub-type="collection"><year>2025</year></pub-date><pub-date pub-type="epub"><day>29</day><month>4</month><year>2025</year></pub-date><volume>4</volume><elocation-id>e72109</elocation-id><history><date date-type="received"><day>04</day><month>02</month><year>2025</year></date><date date-type="rev-recd"><day>26</day><month>03</month><year>2025</year></date><date date-type="accepted"><day>27</day><month>03</month><year>2025</year></date></history><copyright-statement>&#x00A9; Yosuke Yamagishi, Shouhei Hanaoka, Tomohiro Kikuchi, Takahiro Nakao, Yuta Nakamura, Yukihiro Nomura, Soichiro Miki, Takeharu Yoshikawa, Osamu Abe. Originally published in JMIR AI (<ext-link ext-link-type="uri" xlink:href="https://ai.jmir.org">https://ai.jmir.org</ext-link>), 29.4.2025. </copyright-statement><copyright-year>2025</copyright-year><license license-type="open-access" xlink:href="https://creativecommons.org/licenses/by/4.0/"><p>This is an open-access article distributed under the terms of the Creative Commons Attribution License (<ext-link ext-link-type="uri" xlink:href="https://creativecommons.org/licenses/by/4.0/">https://creativecommons.org/licenses/by/4.0/</ext-link>), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work, first published in JMIR AI, is properly cited. The complete bibliographic information, a link to the original publication on <ext-link ext-link-type="uri" xlink:href="https://www.ai.jmir.org/">https://www.ai.jmir.org/</ext-link>, as well as this copyright and license information must be included.</p></license><self-uri xlink:type="simple" xlink:href="https://ai.jmir.org/2025/1/e72109"/><abstract><sec><title>Background</title><p>Medical image segmentation is crucial for diagnosis and treatment planning in radiology, but it traditionally requires extensive manual effort and specialized training data. With its novel video tracking capabilities, the Segment Anything Model 2 (SAM 2) presents a potential solution for automated 3D medical image segmentation without the need for domain-specific training. However, its effectiveness in medical applications, particularly in abdominal computed tomography (CT) imaging remains unexplored.</p></sec><sec><title>Objective</title><p>The aim of this study was to evaluate the zero-shot performance of SAM 2 in 3D segmentation of abdominal organs in CT scans and to investigate the effects of prompt settings on segmentation results.</p></sec><sec sec-type="methods"><title>Methods</title><p>In this retrospective study, we used a subset of the TotalSegmentator CT dataset from eight institutions to assess SAM 2&#x2019;s ability to segment eight abdominal organs. Segmentation was initiated from three different z-coordinate levels (caudal, mid, and cranial levels) of each organ. Performance was measured using the dice similarity coefficient (DSC). We also analyzed the impact of &#x201C;negative prompts,&#x201D; which explicitly exclude certain regions from the segmentation process, on accuracy.</p></sec><sec sec-type="results"><title>Results</title><p>A total of 123 patients (mean age 60.7, SD 15.5 years; 63 men, 60 women) were evaluated. As a zero-shot approach, larger organs with clear boundaries demonstrated high segmentation performance, with mean DSCs as follows: liver, 0.821 (SD 0.192); right kidney, 0.862 (SD 0.212); left kidney, 0.870 (SD 0.154); and spleen, 0.891 (SD 0.131). Smaller organs showed lower performance: gallbladder, 0.531 (SD 0.291); pancreas, 0.361 (SD 0.197); and adrenal glands&#x2014;right, 0.203 (SD 0.222) and left, 0.308 (SD 0.234). The initial slice for segmentation and the use of negative prompts significantly influenced the results. By removing negative prompts from the input, the DSCs significantly decreased for six organs.</p></sec><sec sec-type="conclusions"><title>Conclusions</title><p>SAM 2 demonstrated promising zero-shot performance in segmenting certain abdominal organs in CT scans, particularly larger organs. Performance was significantly influenced by input negative prompts and initial slice selection, highlighting the importance of optimizing these factors.</p></sec></abstract><kwd-group><kwd>artificial intelligence</kwd><kwd>medical image processing</kwd><kwd>computed tomography</kwd><kwd>abdominal imaging</kwd><kwd>segmentation</kwd><kwd>AI</kwd></kwd-group></article-meta></front><body><sec id="s1" sec-type="intro"><title>Introduction</title><p>Medical image segmentation is a critical task in radiology, playing a vital role in diagnosis, treatment planning, and clinical research [<xref ref-type="bibr" rid="ref1">1</xref>,<xref ref-type="bibr" rid="ref2">2</xref>]. Traditionally, this process has been labor-intensive, requiring manual delineation by skilled radiologists. However, recent advancements in deep learning have revolutionized this field, expanding the scope of automated analysis and significantly enhancing performance across diverse medical imaging tasks.</p><p>The Segment Anything Model (SAM), introduced by Meta AI, represented a significant leap forward in image segmentation technology [<xref ref-type="bibr" rid="ref3">3</xref>]. Trained on over a billion masks, SAM demonstrated remarkable versatility in segmenting a wide array of objects across various domains. SAM&#x2019;s zero-shot performance&#x2014;its ability to segment objects it has never seen during training&#x2014;in medical images has been extensively evaluated [<xref ref-type="bibr" rid="ref4">4</xref>,<xref ref-type="bibr" rid="ref5">5</xref>], and specialized models such as MedSAM [<xref ref-type="bibr" rid="ref6">6</xref>], which underwent additional training for medical imaging applications, have been introduced. These developments have demonstrated SAM&#x2019;s potential in radiological domains, including CT and magnetic resonance imaging (MRI). However, SAM was primarily designed for 2D image segmentation, which imposed inherent limitations on its direct applicability to 3D volumetric data.</p><p>SAM 2, released in July 2024, introduced video segmentation capabilities [<xref ref-type="bibr" rid="ref7">7</xref>], applicable to 3D medical imaging like CT scans. Although not specifically designed for medical use, its zero-shot ability and video tracking features offer a promising approach to 3D medical image segmentation, potentially overcoming limitations of traditional methods that require extensive domain-specific training. Testing SAM 2&#x2019;s zero-shot performance is crucial because it could significantly reduce the need for large, annotated medical datasets and specialized model training, potentially accelerating the deployment of artificial intelligence in various medical imaging applications.</p><p>SAM 2&#x2019;s zero-shot performance in radiology and the impact of input factors remain understudied, despite evaluations in surgical video segmentation [<xref ref-type="bibr" rid="ref8">8</xref>] and specialized versions like Medical SAM2 [<xref ref-type="bibr" rid="ref9">9</xref>]. We assess SAM 2&#x2019;s zero-shot performance in medical imaging, examining how target organ size, initial slice selection, and negative prompts influence its segmentation accuracy. These factors are crucial for optimizing SAM 2&#x2019;s performance in radiological applications.</p><p>We focused our evaluation on abdominal organs due to their significant clinical importance. Morphological and size analysis of these organs is crucial for disease detection [<xref ref-type="bibr" rid="ref10">10</xref>]; pancreatic atrophy may indicate highly fatal pancreatic cancer [<xref ref-type="bibr" rid="ref11">11</xref>] and liver morphology changes can signal cirrhosis [<xref ref-type="bibr" rid="ref12">12</xref>]. Renal atrophy is associated with chronic kidney diseases [<xref ref-type="bibr" rid="ref13">13</xref>], and a recent study has shown that kidney volume measurements obtained through accurate segmentation models effectively predict kidney function [<xref ref-type="bibr" rid="ref14">14</xref>]. These volumetric assessments require precise 3D segmentation, making abdominal imaging an ideal test case for evaluating SAM 2&#x2019;s capabilities in clinically relevant scenarios.</p><p>This comprehensive evaluation combining zero-shot performance assessment and input factor analysis is one of the earliest investigations for SAM 2 applied to 3D medical imaging. Our exploration is analogous to large language models, where performance varies significantly based on prompt adjustments [<xref ref-type="bibr" rid="ref15">15</xref>]. By examining prompt engineering in segmentation, we aim to provide deeper insights into adapting general-purpose AI models for specialized medical imaging applications.</p></sec><sec id="s2" sec-type="methods"><title>Methods</title><p>This study was conducted as a retrospective study and adheres to the Checklist for Artificial Intelligence in Medical Imaging (CLAIM): 2024 Update [<xref ref-type="bibr" rid="ref16">16</xref>,<xref ref-type="bibr" rid="ref17">17</xref>].</p><sec id="s2-1"><title>Ethical Considerations</title><p>We used the openly available TotalSegmentator dataset [18], a CT image segmentation dataset. The TotalSegmentator dataset is released under the Creative Commons Attribution 4.0 International license, permitting unrestricted reuse for research. The original CT images were collected retrospectively at University Hospital Basel. The Ethics Committee Northwest and Central Switzerland (EKNZ) approved a waiver of ethical approval for that retrospective study (BASEC Req-2022&#x2013;00495). No additional ethics review was required for this secondary analysis. Patient consent was waived by the EKNZ due to the deidentified, retrospective nature of the original data collection. All CT images in the TotalSegmentator dataset were fully deidentified of any protected health information before public release. The dataset contains anonymized images and no patient identifiers. No financial or other compensation was provided to participants for the original data collection, and none was provided for this secondary analysis.</p></sec><sec id="s2-2"><title>Dataset</title><p>We aimed to evaluate the segmentation performance of major organs within the imaging range of abdominal CT, one of the most common medical imaging modalities. To conduct this performance evaluation, we used a subset of the TotalSegmentator CT dataset version 1.0 [<xref ref-type="bibr" rid="ref18">18</xref>]. The TotalSegmentator dataset is a large-scale, multiorgan segmentation dataset collected from eight institutions. We selected this dataset for its comprehensive organ segmentation masks and available institutional metadata for each case. These segmentation masks underwent expert verification to ensure high quality, making the dataset particularly suitable for our study.</p><p>Our study included cases that encompassed the abdominal region, while CT angiography scans were excluded from the analysis.</p><p>To ensure representation from all eight institutions while managing the dataset size, we implemented a sampling strategy. We set a maximum of 20 cases per institution and randomly selected cases up to this limit. For institutions with fewer than 20 cases, all available cases were included.</p><p>We focused on 8 major abdominal organs for our analysis:</p><list list-type="order"><list-item><p>Liver</p></list-item><list-item><p>Right kidney</p></list-item><list-item><p>Left kidney</p></list-item><list-item><p>Spleen</p></list-item><list-item><p>Gallbladder</p></list-item><list-item><p>Pancreas</p></list-item><list-item><p>Right adrenal gland</p></list-item><list-item><p>Left adrenal gland</p></list-item></list><p>These organs were selected based on their clinical significance and visibility in standard abdominal CT scans. To account for potential annotation deficiencies, we excluded segmentation masks with extremely small volumes by setting a threshold of 100 voxels. Masks below this threshold were omitted from the analysis. The dataset selection flowchart is illustrated in <xref ref-type="fig" rid="figure1">Figure 1</xref>.</p><fig position="float" id="figure1"><label>Figure 1.</label><caption><p>Flow diagram illustrating the CT scan selection process from the TotalSegmentator dataset for evaluation of SAM 2. CT: computed tomography; SAM 2: Segment Anything Model 2.</p></caption><graphic alt-version="no" mimetype="image" position="float" xlink:type="simple" xlink:href="ai_v4i1e72109_fig01.png"/></fig></sec><sec id="s2-3"><title>Data Preprocessing</title><p>The dataset was available in NIfTI file format. For SAM 2 inference, we extracted each horizontal slice from the 3D volumes to create subsets of 2D images for each scan. We applied windowing to the CT scans, using a window level of 50 and a window width of 400 Hounsfield units. Following windowing, we performed min-max scaling on the data. The scaled values were then converted to 8-bit integers, resulting in a range of 0&#x2010;255. These processed 2D images were saved as sequential JPEG files.</p><p>For SAM 2 inference, we selectively processed only the slices containing abdominal organs. This approach focused on optimizing computational efficiency, resulting in faster inference speeds.</p></sec><sec id="s2-4"><title>Analysis of Organ Mask Volumes</title><p>For the volumetric analysis, we used the existing segmentation masks from the TotalSegmentator dataset to calculate the volume of each organ in voxels. We chose to measure in voxels rather than physical units, as our model inputs do not consider voxel scale.</p><p>Additionally, we analyzed cross-sectional areas of organ masks along the z-axis. For each organ, we calculated mean mask areas at the 25th, 50th, and 75th percentile z-coordinates, corresponding to the initial prompt locations used in SAM 2 segmentation.</p></sec><sec id="s2-5"><title>SAM 2 Implementation for 3D Medical Image Segmentation</title><p>SAM 2 is a segmentation model not specifically designed for medical images, but for general video content such as sports or animal footage. These models are trained on a large-scale dataset, enabling them to perform segmentation on any object. SAM 2&#x2019;s main feature is its ability to support not only 2D images but also videos. By providing coordinates indicating the target object for segmentation, SAM 2 can track and segment objects appearing within the video. An overview of CT volume segmentation using SAM 2&#x2019;s video predictor is shown in <xref ref-type="fig" rid="figure2">Figure 2</xref>.</p><fig position="float" id="figure2"><label>Figure 2.</label><caption><p>Workflow of 3D medical image segmentation using SAM 2. This figure illustrates a two-stage process for 3D medical image segmentation. The top row shows the prompt input process, where a slice is selected from a CT volume as an initial frame and XY-coordinates are added as prompts for organ segmentation (red crosses: positive prompts; blue crosses: negative prompts). These are then input into SAM 2. The bottom row depicts the 3D segmentation process, where SAM 2 performs forward and reverse predictions to generate 3D segmentation masks at both cranial and caudal sides. These masks are ultimately merged to create a full 3D segmentation mask of the target organ. CT: computed tomography; SAM 2: Segment Anything Model 2.</p></caption><graphic alt-version="no" mimetype="image" position="float" xlink:type="simple" xlink:href="ai_v4i1e72109_fig02.png"/></fig><sec id="s2-5-1"><title>Adaptation for 3D Medical Imaging</title><p>Although SAM 2 was originally intended for tracking objects in general videos, we recognized that 3D volumes such as CT and MRI scans can be considered as videos composed of numerous 2D images. Using publicly available datasets with segmentation masks, we applied preprocessing compatible with SAM 2&#x2019;s video prediction capabilities. This allowed us to construct a pipeline capable of performing multiorgan segmentation in a zero-shot manner, without additional training of SAM 2.</p></sec><sec id="s2-5-2"><title>Bidirectional Prediction Approach</title><p>While SAM 2&#x2019;s video prediction is unidirectional, it can process in both directions from the initial frame. To obtain a complete segmentation mask for the entire volume, we implemented a simple bidirectional approach: forward direction from the starting slice to the cranial end reverse direction from the starting slice to the caudal end. The two segmentation masks obtained from these bidirectional inferences were then merged to create a complete 3D segmentation mask.</p></sec><sec id="s2-5-3"><title>Model Inference and Prompt Setting</title><p>SAM 2&#x2019;s video prediction requires input of both the video (numbered 2D images) and prompt (coordinates for the target object). We used axial slices for prompt input, as these views serve as the foundation of radiological interpretation in clinical practice. This approach aligns with the standard workflow of radiologists when adapting SAM2 for medical image analysis. In practice, the prompt must be manually specified by a user. However, given the need to evaluate a large number of objects, we devised an algorithm to automatically obtain prompts:</p><list list-type="order"><list-item><p>Z-coordinate focus: Using 25th (caudal-level), 50th (mid-level), and 75th percentiles (cranial-level) for comprehensive organ representation.</p></list-item><list-item><p>Random selection within organ boundaries:</p><list list-type="bullet"><list-item><p>Five positive prompts from within the segmentation mask</p></list-item><list-item><p>Five negative prompts were sampled from regions 2&#x2010;3 voxels outside the mask boundary, excluding the immediate 1-voxel margin</p></list-item></list></list-item></list><p>This method maintains reproducibility, reduces bias from the user&#x2019;s prompting skill, and leverages SAM 2&#x2019;s capability to use both positive and negative prompts for improved accuracy.</p><p>In this study, we refer to the 3D segmentations initiated at each of these positions as caudal-approach, mid-approach, and cranial-approach, corresponding to segmentations starting from the caudal, mid, and cranial-level slices, respectively.</p></sec><sec id="s2-5-4"><title>Model Version</title><p>For our study, we selected the &#x201C;sam2_hiera_large&#x201D; model due to its superior performance among the available versions. We used version 1.0 of the SAM 2 [<xref ref-type="bibr" rid="ref19">19</xref>]. Our implementation was carried out using Python (version 3.10.12).</p></sec></sec><sec id="s2-6"><title>Statistical Analysis</title><p>To evaluate the model&#x2019;s performance, we calculated the Dice similarity coefficient (DSC) [<xref ref-type="bibr" rid="ref20">20</xref>]. This evaluation was performed organ-wise across the dataset to provide a detailed analysis.</p><p>We then compared the segmentation performance across the different approaches with or without negative prompts. We performed three pairwise comparisons for the approaches: caudal-approach versus mid-approach, caudal-approach versus cranial-approach, and mid-approach versus cranial-approach. Additionally, we compared performance with and without negative prompts.</p><p>When considering organ volumes in detail, we calculated Spearman correlation coefficients to examine the relationship between organ volumes and DSCs [<xref ref-type="bibr" rid="ref21">21</xref>,<xref ref-type="bibr" rid="ref22">22</xref>].</p><p>To account for multiple comparisons, we applied the Bonferroni correction. After Bonferroni correction, a P value of &#x003C;.05/3 (approximately 0.0167) was considered statistically significant.</p><p>All statistical analyses were conducted using SciPy (version 1.14.0).</p></sec></sec><sec id="s3" sec-type="results"><title>Results</title><sec id="s3-1"><title>Data Characteristics</title><p>Our sampling strategy resulted in a total of 123 scans. Twenty scans each were selected from five institutions, while the remaining three institutions contributed 5, 5, and 13 cases respectively. The average age of the patients in our selected sample was 60.7 (SD 15.5) years. The gender distribution was nearly equal: 63 male and 60 female individuals.</p><p>903 organ segmentations were obtained from 123 scans. 12 masks with volumes of 100 voxels or smaller were then excluded from the analysis, and the final dataset consisted of 891 organ segmentations.</p></sec><sec id="s3-2"><title>Analysis of Organ Mask Volumes and Areas</title><p>Organ volumes, measured in voxels and are detailed in <xref ref-type="table" rid="table1">Table 1</xref>. The liver was the largest organ, followed by the spleen; kidneys were next in size, with similar volumes on the left and right side. Compared to the liver&#x2019;s mean volume, the pancreas was approximately 1/25th the size of the liver; the gallbladder was less than 1/70, and both adrenal glands were less than 1/400. These three organs (ie, pancreas, gallbladder, and adrenal glands) can be categorized as small organs.</p><table-wrap id="t1" position="float"><label>Table 1.</label><caption><p>Descriptive statistics of organ volumes in voxels derived from computed tomography scan mask volumes.</p></caption><table id="table1" frame="hsides" rules="groups"><thead><tr><td align="left" valign="bottom">Organ</td><td align="left" valign="bottom">Organ volumes (voxels), mean (SD)</td><td align="left" valign="bottom">Min<sup><xref ref-type="table-fn" rid="table1fn1">a</xref></sup></td><td align="left" valign="bottom">Max<sup><xref ref-type="table-fn" rid="table1fn2">b</xref></sup></td><td align="left" valign="bottom">Samples (n)</td></tr></thead><tbody><tr><td align="left" valign="top">Liver</td><td align="left" valign="top">465,008.6 (156,091.00)</td><td align="left" valign="top">19,768</td><td align="left" valign="top">963,401</td><td align="left" valign="top">119</td></tr><tr><td align="left" valign="top">Right kidney</td><td align="left" valign="top">39,381.57 (18,122.20)</td><td align="left" valign="top">216</td><td align="left" valign="top">79,713</td><td align="left" valign="top">108</td></tr><tr><td align="left" valign="top">Left kidney</td><td align="left" valign="top">41,246.74 (21,144.60)</td><td align="left" valign="top">666</td><td align="left" valign="top">129,706</td><td align="left" valign="top">111</td></tr><tr><td align="left" valign="top">Spleen</td><td align="left" valign="top">71,730.34 (45,884.40)</td><td align="left" valign="top">13,818</td><td align="left" valign="top">303,676</td><td align="left" valign="top">115</td></tr><tr><td align="left" valign="top">Gallbladder</td><td align="left" valign="top">6,247.61 (4,902.72)</td><td align="left" valign="top">170</td><td align="left" valign="top">20,763</td><td align="left" valign="top">89</td></tr><tr><td align="left" valign="top">Pancreas</td><td align="left" valign="top">18,526.41 (8,502.56)</td><td align="left" valign="top">707</td><td align="left" valign="top">37,855</td><td align="left" valign="top">116</td></tr><tr><td align="left" valign="top">Right adrenal gland</td><td align="left" valign="top">1,101.86 (465.47)</td><td align="left" valign="top">216</td><td align="left" valign="top">2590</td><td align="left" valign="top">118</td></tr><tr><td align="left" valign="top">Left adrenal gland</td><td align="left" valign="top">1,259.03 (522.81)</td><td align="left" valign="top">135</td><td align="left" valign="top">2977</td><td align="left" valign="top">115</td></tr></tbody></table><table-wrap-foot><fn id="table1fn1"><p><sup>a</sup>Min: Minimum</p></fn><fn id="table1fn2"><p><sup>b</sup>Max: Maximum</p></fn></table-wrap-foot></table-wrap><p>Organ cross-sectional area analysis showed diverse trends across eight organs. The liver was the largest in size, increasing caudally to cranially. The pancreas steadily increased while adrenal glands, though smallest, peaked at mid-level. The details are provided in Figure S1 in <xref ref-type="supplementary-material" rid="app1">Multimedia Appendix 1</xref>.</p></sec><sec id="s3-3"><title>Multiorgan Segmentation Performance</title><p>We evaluated the performance for multiorgan segmentation using different starting slice positions. The DSCs are reported as mean (SD) to reflect performance variability. All results are detailed in <xref ref-type="table" rid="table2">Table 2</xref>.</p><table-wrap id="t2" position="float"><label>Table 2.</label><caption><p>DSCs<sup><xref ref-type="table-fn" rid="table2fn1">a</xref></sup> for multiorgan segmentation by different approaches (ie, caudal, mid, and cranial).</p></caption><table id="table2" frame="hsides" rules="groups"><thead><tr><td align="left" valign="bottom">Organ and approach</td><td align="left" valign="bottom">DSC,<sup><xref ref-type="table-fn" rid="table2fn1">a</xref></sup> mean (SD)</td><td align="left" valign="bottom"><italic>P</italic> value<sup><xref ref-type="table-fn" rid="table2fn2">b</xref></sup></td></tr></thead><tbody><tr><td align="left" valign="top">Liver</td><td align="left" valign="top"/><td align="left" valign="top">caudal versus mid: &#x003C;.01 caudal versus cranial: &#x003C;.01mid versus cranial: .07</td></tr><tr><td align="left" valign="top">&#x2003;caudal</td><td align="char" char="." valign="top">0.821 (0.192)</td><td align="left" valign="top"/></tr><tr><td align="left" valign="top">&#x2003;mid</td><td align="char" char="." valign="top">0.754 (0.223)</td><td align="left" valign="top"/></tr><tr><td align="left" valign="top">&#x2003;cranial</td><td align="char" char="." valign="top">0.702 (0.259)</td><td align="left" valign="top"/></tr><tr><td align="left" valign="top">Right kidney</td><td align="left" valign="top"/><td align="left" valign="top">caudal versus mid: .03 caudal versus cranial: .16 mid versus cranial: &#x003C;.01</td></tr><tr><td align="left" valign="top">&#x2003;caudal</td><td align="char" char="." valign="top">0.862 (0.189)</td><td align="left" valign="top"/></tr><tr><td align="left" valign="top">&#x2003;mid</td><td align="char" char="." valign="top">0.862 (0.212)</td><td align="left" valign="top"/></tr><tr><td align="left" valign="top">&#x2003;cranial</td><td align="char" char="." valign="top">0.801 (0.270)</td><td align="left" valign="top"/></tr><tr><td align="left" valign="top">Left kidney</td><td align="left" valign="top"/><td align="left" valign="top">caudal versus mid: .40 caudal versus cranial: .15 mid versus cranial: .15</td></tr><tr><td align="left" valign="top">&#x2003;caudal</td><td align="char" char="." valign="top">0.870 (0.154)</td><td align="left" valign="top"/></tr><tr><td align="left" valign="top">&#x2003;mid</td><td align="left" valign="top">0.825 (0.221)</td><td align="left" valign="top"/></tr><tr><td align="left" valign="top">&#x2003;cranial</td><td align="left" valign="top">0.808 (0.242)</td><td align="left" valign="top"/></tr><tr><td align="left" valign="top">Spleen</td><td align="left" valign="top"/><td align="left" valign="top">caudal versus mid: &#x003C;.01 caudal versus cranial: .017 mid versus cranial: .56</td></tr><tr><td align="left" valign="top">&#x2003;caudal</td><td align="char" char="." valign="top">0.891 (0.131)</td><td align="left" valign="top"/></tr><tr><td align="left" valign="top">&#x2003;mid</td><td align="left" valign="top">0.839 (0.187)</td><td align="left" valign="top"/></tr><tr><td align="left" valign="top">&#x2003;cranial</td><td align="left" valign="top">0.768 (0.302)</td><td align="left" valign="top"/></tr><tr><td align="left" valign="top">Gallbladder</td><td align="left" valign="top"/><td align="left" valign="top">caudal versus mid: .95 caudal versus cranial: &#x003C;.01mid versus cranial: .08</td></tr><tr><td align="left" valign="top">&#x2003;caudal</td><td align="char" char="." valign="top">0.527 (0.288)</td><td align="left" valign="top"/></tr><tr><td align="left" valign="top">&#x2003;mid</td><td align="left" valign="top">0.531 (0.291)</td><td align="left" valign="top"/></tr><tr><td align="left" valign="top">&#x2003;cranial</td><td align="left" valign="top">0.461 (0.314)</td><td align="left" valign="top"/></tr><tr><td align="left" valign="top">Pancreas</td><td align="left" valign="top"/><td align="left" valign="top">caudal versus mid: .92 caudal versus cranial: &#x003C;.01mid versus cranial: &#x003C;.01</td></tr><tr><td align="left" valign="top">&#x2003;caudal</td><td align="char" char="." valign="top">0.353 (0.168)</td><td align="left" valign="top"/></tr><tr><td align="left" valign="top">&#x2003;mid</td><td align="left" valign="top">0.361 (0.197)</td><td align="left" valign="top"/></tr><tr><td align="left" valign="top">&#x2003;cranial</td><td align="left" valign="top">0.287 (0.209)</td><td align="left" valign="top"/></tr><tr><td align="left" valign="top">Right adrenal gland</td><td align="left" valign="top"/><td align="left" valign="top">caudal versus mid: &#x003C;.01 caudal versus cranial: &#x003C;.01mid versus cranial: &#x003C;.01</td></tr><tr><td align="left" valign="top">&#x2003;caudal</td><td align="char" char="." valign="top">0.203 (0.222)</td><td align="left" valign="top"/></tr><tr><td align="left" valign="top">&#x2003;mid</td><td align="left" valign="top">0.177 (0.235)</td><td align="left" valign="top"/></tr><tr><td align="left" valign="top">&#x2003;cranial</td><td align="left" valign="top">0.112 (0.178)</td><td align="left" valign="top"/></tr><tr><td align="left" valign="top">Left adrenal gland</td><td align="left" valign="top"/><td align="left" valign="top">caudal versus mid: &#x003C;.01 caudal versus cranial: &#x003C;.01mid versus cranial: .08</td></tr><tr><td align="left" valign="top">&#x2003;caudal</td><td align="char" char="." valign="top">0.308 (0.234)</td><td align="left" valign="top"/></tr><tr><td align="left" valign="top">&#x2003;mid</td><td align="left" valign="top">0.252 (0.226)</td><td align="left" valign="top"/></tr><tr><td align="left" valign="top">&#x2003;cranial</td><td align="left" valign="top">0.226 (0.238)</td><td align="left" valign="top"/></tr></tbody></table><table-wrap-foot><fn id="table2fn1"><p><sup>a</sup>DSC: Dice Similarity Coefficient.</p></fn><fn id="table2fn2"><p><sup>b</sup><italic>P</italic> values from Wilcoxon signed-rank tests are provided for comparisons between approaches (caudal vs mid, caudal vs cranial, and mid vs cranial).</p></fn></table-wrap-foot></table-wrap><p>The left kidney demonstrated the best overall performance, maintaining high DSCs across all starting positions: mean 0.870 (SD 0.154) , mean 0.825 (0.221), and 0.808 (SD 0.242) for the caudal-approach, mid-approach, and cranial-approach, respectively. Notably, it was the only organ showing no statistically significant differences between any starting positions (all <italic>P</italic>&#x003E;.0167).</p><p>The box plots (<xref ref-type="fig" rid="figure3">Figure 3</xref>) show that most organs reached DSCs above 0.8, with some approaching or nearly reaching 1.0. However, the box plots also reveal instances of very low DSC values approaching 0 across various organs and approaches, indicating significant variability in segmentation performance.</p><p>Starting slice level had a significant impact on most organs. Organs demonstrated various patterns in segmentation performance depending on the starting level. The liver showed a significant decrease in performance as the starting position moved superiorly, with DSC dropping from mean 0.821 (SD 0.192) with mean caudal-approach to 0.702 (SD 0.259) with cranial-approach (<italic>P&#x003C;</italic>.01). Smaller organs such as the pancreas, adrenal glands, and gallbladder showed the most pronounced impact of the starting position. For these organs, performance significantly decreased when changing from a caudal-approach to a cranial-approach (all <italic>Ps&#x003C;.</italic>01).</p><p>Larger organs such as liver, kidneys, and spleen consistently demonstrated higher DSCs compared to smaller organs across all approaches. A moderate correlation was observed across all settings when using caudal-approach (&#x03C1;=0.731; <italic>P&#x003C;.</italic>01), mid-approach (&#x03C1;=0.698; <italic>P&#x003C;.</italic>01), and cranial-approach (&#x03C1;=0.699; <italic>P&#x003C;.</italic>01) (12,13). As shown in <xref ref-type="table" rid="table3">Table 3</xref>, when correlation coefficients were calculated separately for each organ, fair correlations were demonstrated for almost all items, particularly for smaller organs.</p><p>We also investigated the impact of including negative prompts on segmentation performance across different organs, focusing specifically on the caudal-approach (<xref ref-type="table" rid="table4">Table 4</xref> and Figure S2 in <xref ref-type="supplementary-material" rid="app1">Multimedia Appendix 1</xref>). All organs except the liver (<italic>P</italic>=.32) and spleen (<italic>P</italic>=.27) demonstrated significant increases in DSC (<italic>P&#x003C;.</italic>01) with the inclusion of negative prompts.</p><fig position="float" id="figure3"><label>Figure 3.</label><caption><p>Box plots of DSCs for eight organs (displayed in separate subplots) across three approaches: caudal-approach, mid-approach, and cranial-approach. Each subplot shows the distribution of DSCs (y-axis, range 0&#x2010;1) for a specific organ, with the three approaches compared along the x-axis. DSC: Dice Similarity Coefficient.</p></caption><graphic alt-version="no" mimetype="image" position="float" xlink:type="simple" xlink:href="ai_v4i1e72109_fig03.png"/></fig><table-wrap id="t3" position="float"><label>Table 3.</label><caption><p>Spearman correlation coefficients between ground truth values of organ volumes and dice similarity coefficients in caudal, mid, and cranial levels.</p></caption><table id="table3" frame="hsides" rules="groups"><thead><tr><td align="left" valign="bottom">Organ</td><td align="left" valign="bottom">Spearman correlation coefficient, &#x03C1; (caudal) &#x03C1;<sup><xref ref-type="table-fn" rid="table3fn1">a</xref></sup></td><td align="left" valign="bottom"><italic>P</italic> value</td><td align="left" valign="bottom">Spearman correlation coefficient, &#x03C1; (mid)<sup><xref ref-type="table-fn" rid="table3fn1">a</xref></sup></td><td align="left" valign="bottom"><italic>P</italic> value</td><td align="left" valign="bottom">Spearman correlation coefficient, &#x03C1; (cranial)<sup><xref ref-type="table-fn" rid="table3fn1">a</xref></sup></td><td align="left" valign="bottom"><italic>P</italic> value</td></tr></thead><tbody><tr><td align="left" valign="top">Liver</td><td align="left" valign="top">0.328</td><td align="char" char="." valign="top">&#x003C;.01</td><td align="left" valign="top">0.0489</td><td align="char" char="." valign="top">.60</td><td align="left" valign="top">0.163</td><td align="char" char="." valign="top">.08</td></tr><tr><td align="left" valign="top">Right kidney</td><td align="left" valign="top">0.231</td><td align="char" char="." valign="top">.02</td><td align="left" valign="top">0.347</td><td align="char" char="." valign="top">&#x003C;.01</td><td align="left" valign="top">0.509</td><td align="char" char="." valign="top">&#x003C;.01</td></tr><tr><td align="left" valign="top">Left kidney</td><td align="left" valign="top">&#x2212;0.0107</td><td align="char" char="." valign="top">.91</td><td align="left" valign="top">0.293</td><td align="char" char="." valign="top">&#x003C;.01</td><td align="left" valign="top">0.295</td><td align="char" char="." valign="top">&#x003C;.01</td></tr><tr><td align="left" valign="top">Spleen</td><td align="left" valign="top">0.38</td><td align="char" char="." valign="top">&#x003C;.01</td><td align="left" valign="top">0.307</td><td align="char" char="." valign="top">&#x003C;.01</td><td align="left" valign="top">0.355</td><td align="char" char="." valign="top">&#x003C;.01</td></tr><tr><td align="left" valign="top">Gallbladder</td><td align="left" valign="top">0.499</td><td align="char" char="." valign="top">&#x003C;.01</td><td align="left" valign="top">0.509</td><td align="char" char="." valign="top">&#x003C;.01</td><td align="left" valign="top">0.469</td><td align="char" char="." valign="top">&#x003C;.01</td></tr><tr><td align="left" valign="top">Pancreas</td><td align="left" valign="top">0.475</td><td align="char" char="." valign="top">&#x003C;.01</td><td align="left" valign="top">0.386</td><td align="char" char="." valign="top">&#x003C;.01</td><td align="left" valign="top">0.379</td><td align="char" char="." valign="top">&#x003C;.01</td></tr><tr><td align="left" valign="top">Right adrenal gland</td><td align="left" valign="top">0.371</td><td align="char" char="." valign="top">&#x003C;.01</td><td align="left" valign="top">0.424</td><td align="char" char="." valign="top">&#x003C;.01</td><td align="left" valign="top">0.278</td><td align="char" char="." valign="top">&#x003C;.01</td></tr><tr><td align="left" valign="top">Left adrenal gland</td><td align="left" valign="top">0.452</td><td align="char" char="." valign="top">&#x003C;.01</td><td align="left" valign="top">0.339</td><td align="char" char="." valign="top">&#x003C;.01</td><td align="left" valign="top">0.447</td><td align="char" char="." valign="top">&#x003C;.01</td></tr></tbody></table><table-wrap-foot><fn id="table3fn1"><p><sup>a</sup>Spearman rank correlation coefficient (&#x03C1;) was used to examine the relationship between object volumes and dice similarity coefficients.</p></fn></table-wrap-foot></table-wrap><table-wrap id="t4" position="float"><label>Table 4.</label><caption><p>Comparison of multiorgan segmentation performance without negative prompts.</p></caption><table id="table4" frame="hsides" rules="groups"><thead><tr><td align="left" valign="bottom">Organ</td><td align="left" valign="bottom">DSC<sup><xref ref-type="table-fn" rid="table4fn1">a</xref></sup> mean (SD)</td><td align="left" valign="bottom">Difference<sup><xref ref-type="table-fn" rid="table4fn2">b</xref></sup></td><td align="left" valign="bottom"><italic>P</italic> value<sup><xref ref-type="table-fn" rid="table4fn3">c</xref></sup></td></tr></thead><tbody><tr><td align="left" valign="top">Liver</td><td align="left" valign="top">0.785 (0.244)</td><td align="left" valign="top">&#x2212;0.036</td><td align="left" valign="top">.32</td></tr><tr><td align="left" valign="top">Right kidney</td><td align="left" valign="top">0.858 (0.203)</td><td align="left" valign="top">&#x2212;0.004</td><td align="left" valign="top">&#x003C;.01</td></tr><tr><td align="left" valign="top">Left kidney</td><td align="left" valign="top">0.847 (0.192)</td><td align="left" valign="top">&#x2212;0.023</td><td align="left" valign="top">&#x003C;.01</td></tr><tr><td align="left" valign="top">Spleen</td><td align="left" valign="top">0.867 (0.213)</td><td align="left" valign="top">&#x2212;0.024</td><td align="left" valign="top">.27</td></tr><tr><td align="left" valign="top">Gallbladder</td><td align="left" valign="top">0.438 (0.338)</td><td align="left" valign="top">&#x2212;0.089</td><td align="left" valign="top">&#x003C;.01</td></tr><tr><td align="left" valign="top">Pancreas</td><td align="left" valign="top">0.277 (0.197)</td><td align="left" valign="top">&#x2212;0.076</td><td align="left" valign="top">&#x003C;.01</td></tr><tr><td align="left" valign="top">Right adrenal gland</td><td align="left" valign="top">0.084 (0.151)</td><td align="left" valign="top">&#x2212;0.119</td><td align="left" valign="top">&#x003C;.01</td></tr><tr><td align="left" valign="top">Left adrenal gland</td><td align="left" valign="top">0.190 (0.230)</td><td align="left" valign="top">&#x2212;0.118</td><td align="left" valign="top">&#x003C;.01</td></tr></tbody></table><table-wrap-foot><fn id="table4fn1"><p><sup>a</sup>DSC: dice similarity coefficient.</p></fn><fn id="table4fn2"><p><sup>b</sup>Change when negative prompts are excluded (negative values indicate lower performance without prompts).</p></fn><fn id="table4fn3"><p><sup>c</sup><italic>P</italic> value represents the results of Wilcoxon signed-rank tests comparing performance with, and without negative prompts for each organ.</p></fn></table-wrap-foot></table-wrap><p>In <xref ref-type="fig" rid="figure4">Figure 4</xref>, we present the highest DSC masks, excluding cases where the ground truth segmentations were incomplete. The highest performing masks, as visualized, generated for each organ in 3D were nearly indistinguishable from the ground truth.</p><fig position="float" id="figure4"><label>Figure 4.</label><caption><p>Successful segmentation results for eight abdominal organs. Each row shows different organ with ground truth (left) and predicted (right) 3D masks. Values in parentheses indicate the DSC for each segmentation. DSC: dice similarity coefficient.</p></caption><graphic alt-version="no" mimetype="image" position="float" xlink:type="simple" xlink:href="ai_v4i1e72109_fig04.png"/></fig><p>On the other hand, there were cases where the performance fluctuated significantly due to differences in the approach. We present an example of the liver segmentation results in <xref ref-type="fig" rid="figure5">Figure 5</xref>. The DSC decreased by 0.564 (from 0.924 with caudal-approach to 0.360 with cranial-approach). For the caudal-approach, the initial slice segmentation appears to have been easier due to clear contrast with the surroundings. The cranial-approach resulted in lower DSC values compared to the caudal-approach. Visual inspection of the segmentation results showed incomplete masking and potential inclusion of the inferior vena cava in the liver mask, particularly in areas where boundaries between the liver, inferior vena cava, and abdominal wall were less distinct.</p><fig position="float" id="figure5"><label>Figure 5.</label><caption><p>Comparison of segmentation results using caudal-approach and cranial-approach, showing 2D axial slices with ground truth and predicted mask for both initial slices (top row, with yellow representing the ground truth of liver, blue and red points indicating negative and positive prompts, respectively), alongside 3D renderings of liver segmentation for ground truth, caudal-approach, and cranial-approach (bottom row).</p></caption><graphic alt-version="no" mimetype="image" position="float" xlink:type="simple" xlink:href="ai_v4i1e72109_fig05.png"/></fig></sec></sec><sec id="s4" sec-type="discussion"><title>Discussion</title><p>To our knowledge, this is the first research that not only validates the performance of zero-shot SAM 2 on abdominal organs but also considers the impact of prompt input strategies such as slice positioning and negative prompts. Our findings demonstrate the potential of SAM 2, a general-purpose segmentation model in segmenting abdominal organs from CT scans. SAM2 showed promising performance for larger organs with clear boundaries, such as the liver, kidneys, and spleen, achieving a mean DSC of 0.821&#x2010;0.891. Although SAM 2 was not specifically designed for medical image analysis, its notable performance suggests potential applicability to a wide range of organs and lesions. The choice of initial prompt position had a significant impact on segmentation accuracy, and the optimal position depended on the organ. Excluding negative prompts led to a significant decrease in DSC for all organs except the spleen and liver, highlighting their importance in segmentation accuracy. SAM 2 struggled with smaller and less defined structures such as the adrenal glands, pancreas, and gallbladder, resulting in lower DSCs. Interestingly, we observed a moderate positive correlation between organ volume and DSCs (&#x03C1;=0.731, <italic>P</italic>&#x003C;.01), suggesting that volume size is one of several key factors influencing segmentation accuracy.</p><p>While prior studies have explored zero-shot segmentation performance in CT and MRI [<xref ref-type="bibr" rid="ref23">23</xref>,<xref ref-type="bibr" rid="ref24">24</xref>], our research makes several unique contributions to this emerging field. Ma et al [<xref ref-type="bibr" rid="ref23">23</xref>] conducted a comprehensive benchmarking of SAM 2 across multiple medical image modalities and demonstrated its potential for transfer learning in the medical domain. Similarly, Dong et al [<xref ref-type="bibr" rid="ref24">24</xref>] explored various prompt strategies and propagation directions for 3D segmentation. In contrast, our work specifically examines abdominal CT imaging, analyzing how prompt positioning and negative prompts significantly influence segmentation outcomes. Unlike previous studies, we investigated segmentation from clinically relevant positions (ie, caudal, mid, and cranial) and found that optimal starting positions vary by organ, with negative prompts being crucial for smaller organ segmentation.</p><p>A key advantage of SAM 2 is its ability to generate segmentation masks with just a few clicks on a single slice, drastically reducing the workload for radiologists who previously relied on labor-intensive manual annotations. Furthermore, optimizing prompt input strategies is essential for achieving even greater model performance, as evidenced by SAM&#x2019;s history of various prompt optimization techniques, including automatic prompt generation and learnable prompts [<xref ref-type="bibr" rid="ref25">25</xref>]. Although the scores are lower compared to previous supervised methods, which can achieve mean DSCs in the upper 0.9 range for some organs, they are still notably high for a zero-shot prediction. Moreover, the ability to segment an entire 3D volume by simply selecting and clicking on a target structure in a single slice is particularly significant. This aligns with challenges typically observed in abdominal organ segmentation, even with supervised 3D models. Notably, supervised approaches like TotalSegmentator (based on nnUNet [<xref ref-type="bibr" rid="ref26">26</xref>]), UNet [<xref ref-type="bibr" rid="ref27">27</xref>], SegUNet [<xref ref-type="bibr" rid="ref28">28</xref>], and SwinUNETR [<xref ref-type="bibr" rid="ref29">29</xref>] also tend to show lower DSC for bilateral adrenal glands and gallbladder compared to other organs [<xref ref-type="bibr" rid="ref30">30</xref>], a trend mirrored in SAM 2&#x2019;s performance. Segmentation performance can be inferred to depend on multiple factors related to the 3D morphology, volume size, and contrast with surrounding tissues of target structure. These findings suggest the importance of optimizing prompts taking into account the characteristics of the targeted structure.</p><p>Our study had several limitations. First, our validation relied on a single dataset of abdominal CT scans, despite being a multi-institutional study. For studies focused on abdominal organs, there are publicly available datasets such as AbdomenCT-1K [<xref ref-type="bibr" rid="ref31">31</xref>], which is included in AbdomenAtlas [<xref ref-type="bibr" rid="ref32">32</xref>,<xref ref-type="bibr" rid="ref33">33</xref>]. To expand the validation to other anatomical structures and imaging modalities, datasets such as Vertebral Segmentation [<xref ref-type="bibr" rid="ref34">34</xref>], TotalSegmentator&#x2019;s MRI [<xref ref-type="bibr" rid="ref35">35</xref>] and Duke Liver datasets [<xref ref-type="bibr" rid="ref36">36</xref>] could also be considered, all of which include segmentation masks for their respective targets. Expanding our validation using these resources would allow for a more robust evaluation. Additionally, as our approach was designed to address zero-shot performance validation, we did not perform any additional training such as fine-tuning. Performance improvements can be expected by using task-specific supervised methods instead of zero-shot. Furthermore, while we used an automated approach to evaluate a large number of organs, there is potential for improved accuracy through manual prompts inputting.</p><p>In conclusion, SAM 2 has demonstrated promising zero-shot performance in segmenting certain abdominal organs in CT scans, particularly larger organs with clear boundaries, highlighting its potential for cross-domain generalization in medical imaging. However, further improvements are needed for smaller and less distinct structures. Our study underscores the importance of applying general models to unseen medical images and optimizing input prompts, which together could significantly enhance the accuracy of medical image segmentation.</p></sec></body><back><notes><sec><title>Data Availability</title><p>The datasets generated or analyzed during this study are available in the Zenodo repository [<xref ref-type="bibr" rid="ref37">37</xref>].</p></sec></notes><fn-group><fn fn-type="conflict"><p>None declared.</p></fn></fn-group><glossary><title>Abbreviations:</title><def-list><def-item><term id="abb1">CLAIM</term><def><p>Checklist for Artificial Intelligence in Medical Imaging</p></def></def-item><def-item><term id="abb2">CT</term><def><p>computed tomography</p></def></def-item><def-item><term id="abb3">DSC</term><def><p>dice similarity coefficient</p></def></def-item><def-item><term id="abb4">EKNZ</term><def><p>Ethics Committee Northwest and Central Switzerland</p></def></def-item><def-item><term id="abb5">MRI</term><def><p>magnetic resonance imaging</p></def></def-item><def-item><term id="abb6">SAM</term><def><p>Segment Anything Model</p></def></def-item></def-list></glossary><ref-list><title>References</title><ref id="ref1"><label>1</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Wang</surname><given-names>S</given-names> </name><name name-style="western"><surname>Summers</surname><given-names>RM</given-names> </name></person-group><article-title>Machine learning and radiology</article-title><source>Med Image Anal</source><year>2012</year><month>07</month><volume>16</volume><issue>5</issue><fpage>933</fpage><lpage>951</lpage><pub-id pub-id-type="doi">10.1016/j.media.2012.02.005</pub-id><pub-id pub-id-type="medline">22465077</pub-id></nlm-citation></ref><ref id="ref2"><label>2</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Saba</surname><given-names>L</given-names> </name><name name-style="western"><surname>Biswas</surname><given-names>M</given-names> </name><name name-style="western"><surname>Kuppili</surname><given-names>V</given-names> </name><etal/></person-group><article-title>The present and future of deep learning in radiology</article-title><source>Eur J Radiol</source><year>2019</year><month>05</month><volume>114</volume><fpage>14</fpage><lpage>24</lpage><pub-id pub-id-type="doi">10.1016/j.ejrad.2019.02.038</pub-id></nlm-citation></ref><ref id="ref3"><label>3</label><nlm-citation citation-type="preprint"><person-group person-group-type="author"><name name-style="western"><surname>Kirillov</surname><given-names>A</given-names> </name><name name-style="western"><surname>Mintun</surname><given-names>E</given-names> </name><name name-style="western"><surname>Ravi</surname><given-names>N</given-names> </name><etal/></person-group><article-title>Segment anything</article-title><source>arXiv</source><comment>Preprint posted online on  Apr 5, 2023</comment><pub-id pub-id-type="doi">10.1109/ICCV51070.2023.00371</pub-id></nlm-citation></ref><ref id="ref4"><label>4</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Mazurowski</surname><given-names>MA</given-names> </name><name name-style="western"><surname>Dong</surname><given-names>H</given-names> </name><name name-style="western"><surname>Gu</surname><given-names>H</given-names> </name><name name-style="western"><surname>Yang</surname><given-names>J</given-names> </name><name name-style="western"><surname>Konz</surname><given-names>N</given-names> </name><name name-style="western"><surname>Zhang</surname><given-names>Y</given-names> </name></person-group><article-title>Segment anything model for medical image analysis: An experimental study</article-title><source>Med Image Anal</source><year>2023</year><month>10</month><volume>89</volume><fpage>102918</fpage><pub-id pub-id-type="doi">10.1016/j.media.2023.102918</pub-id><pub-id pub-id-type="medline">37595404</pub-id></nlm-citation></ref><ref id="ref5"><label>5</label><nlm-citation citation-type="preprint"><person-group person-group-type="author"><name name-style="western"><surname>Roy</surname><given-names>S</given-names> </name><name name-style="western"><surname>Wald</surname><given-names>T</given-names> </name><name name-style="western"><surname>Koehler</surname><given-names>G</given-names> </name><etal/></person-group><article-title>SAM.MD: Zero-shot medical image segmentation capabilities of the segment anything model</article-title><source>arXiv</source><comment>Preprint posted online on 2023</comment><pub-id pub-id-type="doi">10.48550/arXiv.2304.05396</pub-id></nlm-citation></ref><ref id="ref6"><label>6</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Ma</surname><given-names>J</given-names> </name><name name-style="western"><surname>He</surname><given-names>Y</given-names> </name><name name-style="western"><surname>Li</surname><given-names>F</given-names> </name><name name-style="western"><surname>Han</surname><given-names>L</given-names> </name><name name-style="western"><surname>You</surname><given-names>C</given-names> </name><name name-style="western"><surname>Wang</surname><given-names>B</given-names> </name></person-group><article-title>Segment anything in medical images</article-title><source>Nat Commun</source><year>2024</year><volume>15</volume><issue>1</issue><fpage>654</fpage><pub-id pub-id-type="doi">10.1038/s41467-024-44824-z</pub-id></nlm-citation></ref><ref id="ref7"><label>7</label><nlm-citation citation-type="preprint"><person-group person-group-type="author"><name name-style="western"><surname>Ravi</surname><given-names>N</given-names> </name><name name-style="western"><surname>Gabeur</surname><given-names>V</given-names> </name><name name-style="western"><surname>Hu</surname><given-names>YT</given-names> </name><etal/></person-group><article-title>SAM 2: Segment Anything in images and videos</article-title><source>arXiv</source><access-date>2024-08-02</access-date><comment>Preprint posted online on  Oct 28, 2024</comment><comment><ext-link ext-link-type="uri" xlink:href="http://arxiv.org/abs/2408.00714">http://arxiv.org/abs/2408.00714</ext-link></comment></nlm-citation></ref><ref id="ref8"><label>8</label><nlm-citation citation-type="preprint"><person-group person-group-type="author"><name name-style="western"><surname>Allan</surname><given-names>M</given-names> </name><name name-style="western"><surname>Kondo</surname><given-names>S</given-names> </name><name name-style="western"><surname>Bodenstedt</surname><given-names>S</given-names> </name><etal/></person-group><article-title>2018 robotic scene segmentation challenge</article-title><source>arXiv</source><comment>Preprint posted online on  Aug 3, 2020</comment><pub-id pub-id-type="doi">10.48550/arXiv.2001.11190</pub-id></nlm-citation></ref><ref id="ref9"><label>9</label><nlm-citation citation-type="preprint"><person-group person-group-type="author"><name name-style="western"><surname>Zhu</surname><given-names>J</given-names> </name><name name-style="western"><surname>Qi</surname><given-names>Y</given-names> </name><name name-style="western"><surname>Wu</surname><given-names>J</given-names> </name></person-group><article-title>Medical SAM 2: Segment medical images as video via segment anything model 2</article-title><source>arXiv</source><access-date>2024-08-07</access-date><comment>Preprint posted online on  Dec 4, 2024</comment><comment><ext-link ext-link-type="uri" xlink:href="http://arxiv.org/abs/2408.00874">http://arxiv.org/abs/2408.00874</ext-link></comment></nlm-citation></ref><ref id="ref10"><label>10</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Geraghty</surname><given-names>EM</given-names> </name><name name-style="western"><surname>Boone</surname><given-names>JM</given-names> </name><name name-style="western"><surname>McGahan</surname><given-names>JP</given-names> </name><name name-style="western"><surname>Jain</surname><given-names>K</given-names> </name></person-group><article-title>Normal organ volume assessment from abdominal CT</article-title><source>Abdom Imaging</source><year>2004</year><volume>29</volume><issue>4</issue><fpage>482</fpage><lpage>490</lpage><pub-id pub-id-type="doi">10.1007/s00261-003-0139-2</pub-id><pub-id pub-id-type="medline">15024516</pub-id></nlm-citation></ref><ref id="ref11"><label>11</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Chu</surname><given-names>LC</given-names> </name><name name-style="western"><surname>Goggins</surname><given-names>MG</given-names> </name><name name-style="western"><surname>Fishman</surname><given-names>EK</given-names> </name></person-group><article-title>Diagnosis and detection of pancreatic cancer</article-title><source>Cancer J</source><year>2017</year><volume>23</volume><issue>6</issue><fpage>333</fpage><lpage>342</lpage><pub-id pub-id-type="doi">10.1097/PPO.0000000000000290</pub-id><pub-id pub-id-type="medline">29189329</pub-id></nlm-citation></ref><ref id="ref12"><label>12</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Gin&#x00E8;s</surname><given-names>P</given-names> </name><name name-style="western"><surname>Krag</surname><given-names>A</given-names> </name><name name-style="western"><surname>Abraldes</surname><given-names>JG</given-names> </name><name name-style="western"><surname>Sol&#x00E0;</surname><given-names>E</given-names> </name><name name-style="western"><surname>Fabrellas</surname><given-names>N</given-names> </name><name name-style="western"><surname>Kamath</surname><given-names>PS</given-names> </name></person-group><article-title>Liver cirrhosis</article-title><source>Lancet</source><year>2021</year><month>10</month><volume>398</volume><issue>10308</issue><fpage>1359</fpage><lpage>1376</lpage><pub-id pub-id-type="doi">10.1016/S0140-6736(21)01374-X</pub-id></nlm-citation></ref><ref id="ref13"><label>13</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Romagnani</surname><given-names>P</given-names> </name><name name-style="western"><surname>Remuzzi</surname><given-names>G</given-names> </name><name name-style="western"><surname>Glassock</surname><given-names>R</given-names> </name><etal/></person-group><article-title>Chronic kidney disease</article-title><source>Nat Rev Dis Primers</source><year>2017</year><month>11</month><day>23</day><volume>3</volume><issue>1</issue><fpage>17088</fpage><pub-id pub-id-type="doi">10.1038/nrdp.2017.88</pub-id><pub-id pub-id-type="medline">29168475</pub-id></nlm-citation></ref><ref id="ref14"><label>14</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Steinhelfer</surname><given-names>L</given-names> </name><name name-style="western"><surname>Jungmann</surname><given-names>F</given-names> </name><name name-style="western"><surname>Nickel</surname><given-names>M</given-names> </name><etal/></person-group><article-title>Automated CT measurement of total kidney volume for predicting renal function decline after 177Lu prostate-specific membrane antigen-I&#x0026;T radioligand therapy</article-title><source>Radiology</source><year>2025</year><month>02</month><volume>314</volume><issue>2</issue><fpage>e240427</fpage><pub-id pub-id-type="doi">10.1148/radiol.240427</pub-id><pub-id pub-id-type="medline">39998377</pub-id></nlm-citation></ref><ref id="ref15"><label>15</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Mesk&#x00F3;</surname><given-names>B</given-names> </name></person-group><article-title>Prompt engineering as an important emerging skill for medical professionals: tutorial</article-title><source>J Med Internet Res</source><year>2023</year><month>10</month><day>4</day><volume>25</volume><fpage>e50638</fpage><pub-id pub-id-type="doi">10.2196/50638</pub-id><pub-id pub-id-type="medline">37792434</pub-id></nlm-citation></ref><ref id="ref16"><label>16</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Mongan</surname><given-names>J</given-names> </name><name name-style="western"><surname>Moy</surname><given-names>L</given-names> </name><name name-style="western"><surname>Kahn</surname><given-names>CE</given-names> </name></person-group><article-title>Checklist for artificial intelligence in medical imaging (claim): a guide for authors and reviewers</article-title><source>Radiol Artif Intell</source><year>2020</year><month>03</month><volume>2</volume><issue>2</issue><fpage>e200029</fpage><pub-id pub-id-type="doi">10.1148/ryai.2020200029</pub-id><pub-id pub-id-type="medline">33937821</pub-id></nlm-citation></ref><ref id="ref17"><label>17</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Tejani</surname><given-names>AS</given-names> </name><name name-style="western"><surname>Klontzas</surname><given-names>ME</given-names> </name><name name-style="western"><surname>Gatti</surname><given-names>AA</given-names> </name><etal/></person-group><article-title>Checklist for artificial intelligence in medical imaging (CLAIM): 2024 update</article-title><source>Radiol Artif Intell</source><year>2024</year><month>07</month><volume>6</volume><issue>4</issue><fpage>e240300</fpage><pub-id pub-id-type="doi">10.1148/ryai.240300</pub-id><pub-id pub-id-type="medline">38809149</pub-id></nlm-citation></ref><ref id="ref18"><label>18</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Wasserthal</surname><given-names>J</given-names> </name><name name-style="western"><surname>Breit</surname><given-names>HC</given-names> </name><name name-style="western"><surname>Meyer</surname><given-names>MT</given-names> </name><etal/></person-group><article-title>TotalSegmentator: robust segmentation of 104 anatomic structures in ct images</article-title><source>Radiol Artif Intell</source><year>2023</year><month>09</month><volume>5</volume><issue>5</issue><fpage>e230024</fpage><pub-id pub-id-type="doi">10.1148/ryai.230024</pub-id><pub-id pub-id-type="medline">37795137</pub-id></nlm-citation></ref><ref id="ref19"><label>19</label><nlm-citation citation-type="web"><article-title>Facebookresearch/sam2</article-title><source>GitHub</source><year>2025</year><access-date>2025-02-04</access-date><comment><ext-link ext-link-type="uri" xlink:href="https://github.com/facebookresearch/sam2">https://github.com/facebookresearch/sam2</ext-link></comment></nlm-citation></ref><ref id="ref20"><label>20</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Dice</surname><given-names>LR</given-names> </name></person-group><article-title>Measures of the amount of ecologic association between species</article-title><source>Ecology</source><year>1945</year><month>07</month><volume>26</volume><issue>3</issue><fpage>297</fpage><lpage>302</lpage><pub-id pub-id-type="doi">10.2307/1932409</pub-id></nlm-citation></ref><ref id="ref21"><label>21</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Akoglu</surname><given-names>H</given-names> </name></person-group><article-title>User&#x2019;s guide to correlation coefficients</article-title><source>Turk J Emerg Med</source><year>2018</year><month>09</month><volume>18</volume><issue>3</issue><fpage>91</fpage><lpage>93</lpage><pub-id pub-id-type="doi">10.1016/j.tjem.2018.08.001</pub-id><pub-id pub-id-type="medline">30191186</pub-id></nlm-citation></ref><ref id="ref22"><label>22</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Chan</surname><given-names>YH</given-names> </name></person-group><article-title>Biostatistics 104: correlational analysis</article-title><source>Singapore Med J</source><year>2003</year><month>12</month><access-date>2025-02-04</access-date><volume>44</volume><issue>12</issue><fpage>614</fpage><lpage>619</lpage><comment><ext-link ext-link-type="uri" xlink:href="https://www.sma.org.sg/smj/4412/4412bs1.pdf">https://www.sma.org.sg/smj/4412/4412bs1.pdf</ext-link></comment><pub-id pub-id-type="medline">14770254</pub-id></nlm-citation></ref><ref id="ref23"><label>23</label><nlm-citation citation-type="preprint"><person-group person-group-type="author"><name name-style="western"><surname>Ma</surname><given-names>J</given-names> </name><name name-style="western"><surname>Kim</surname><given-names>S</given-names> </name><name name-style="western"><surname>Li</surname><given-names>F</given-names> </name><etal/></person-group><article-title>Segment anything in medical images and videos: benchmark and deployment</article-title><source>arXiv</source><comment>Preprint posted online on  Aug 6, 2024</comment><pub-id pub-id-type="doi">10.48550/arXiv.2408.03322</pub-id></nlm-citation></ref><ref id="ref24"><label>24</label><nlm-citation citation-type="preprint"><person-group person-group-type="author"><name name-style="western"><surname>Dong</surname><given-names>H</given-names> </name><name name-style="western"><surname>Gu</surname><given-names>H</given-names> </name><name name-style="western"><surname>Chen</surname><given-names>Y</given-names> </name><name name-style="western"><surname>Yang</surname><given-names>J</given-names> </name><name name-style="western"><surname>Chen</surname><given-names>Y</given-names> </name><name name-style="western"><surname>Mazurowski</surname><given-names>MA</given-names> </name></person-group><article-title>Segment anything model 2: an application to 2D and 3D medical images</article-title><source>arXiv</source><comment>Preprint posted online on  Aug 22, 2024</comment><pub-id pub-id-type="doi">10.48550/arXiv.2408.00756</pub-id></nlm-citation></ref><ref id="ref25"><label>25</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Zhang</surname><given-names>Y</given-names> </name><name name-style="western"><surname>Shen</surname><given-names>Z</given-names> </name><name name-style="western"><surname>Jiao</surname><given-names>R</given-names> </name></person-group><article-title>Segment anything model for medical image segmentation: current applications and future directions</article-title><source>Comput Biol Med</source><year>2024</year><month>03</month><volume>171</volume><fpage>108238</fpage><pub-id pub-id-type="doi">10.1016/j.compbiomed.2024.108238</pub-id></nlm-citation></ref><ref id="ref26"><label>26</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Isensee</surname><given-names>F</given-names> </name><name name-style="western"><surname>Jaeger</surname><given-names>PF</given-names> </name><name name-style="western"><surname>Kohl</surname><given-names>SAA</given-names> </name><name name-style="western"><surname>Petersen</surname><given-names>J</given-names> </name><name name-style="western"><surname>Maier-Hein</surname><given-names>KH</given-names> </name></person-group><article-title>nnU-Net: a self-configuring method for deep learning-based biomedical image segmentation</article-title><source>Nat Methods</source><year>2021</year><month>02</month><volume>18</volume><issue>2</issue><fpage>203</fpage><lpage>211</lpage><pub-id pub-id-type="doi">10.1038/s41592-020-01008-z</pub-id><pub-id pub-id-type="medline">33288961</pub-id></nlm-citation></ref><ref id="ref27"><label>27</label><nlm-citation citation-type="book"><person-group person-group-type="author"><name name-style="western"><surname>Ronneberge</surname><given-names>O</given-names> </name><name name-style="western"><surname>Fischer</surname><given-names>P</given-names> </name><name name-style="western"><surname>Brox</surname><given-names>T</given-names> </name></person-group><person-group person-group-type="editor"><name name-style="western"><surname>Navab</surname><given-names>N</given-names> </name><name name-style="western"><surname>Hornegger</surname><given-names>J</given-names> </name><name name-style="western"><surname>Wells</surname><given-names>WM</given-names> </name><name name-style="western"><surname>Frangi</surname><given-names>AF</given-names> </name></person-group><article-title>U-net: convolutional networks for biomedical image segmentation</article-title><source>Medical Image Computing and Computer-Assisted Intervention &#x2013; MICCAI 2015</source><year>2015</year><publisher-name>Springer International Publishing</publisher-name><fpage>234</fpage><lpage>241</lpage><pub-id pub-id-type="doi">10.1007/978-3-319-24574-4_28</pub-id></nlm-citation></ref><ref id="ref28"><label>28</label><nlm-citation citation-type="book"><person-group person-group-type="author"><name name-style="western"><surname>Chen</surname><given-names>X</given-names> </name><name name-style="western"><surname>Cheng</surname><given-names>G</given-names> </name><name name-style="western"><surname>Cai</surname><given-names>Y</given-names> </name><name name-style="western"><surname>Wen</surname><given-names>D</given-names> </name><name name-style="western"><surname>Li</surname><given-names>H</given-names> </name></person-group><person-group person-group-type="editor"><name name-style="western"><surname>Tan</surname><given-names>T</given-names> </name><name name-style="western"><surname>Li</surname><given-names>X</given-names> </name><name name-style="western"><surname>Chen</surname><given-names>X</given-names> </name><name name-style="western"><surname>Zhou</surname><given-names>J</given-names> </name><name name-style="western"><surname>Yang</surname><given-names>J</given-names> </name><name name-style="western"><surname>Cheng</surname><given-names>H</given-names> </name></person-group><article-title>Semantic segmentation with modified deep residual networks</article-title><source>Pattern Recognit DAGM</source><year>2016</year><publisher-name>Springer</publisher-name><fpage>42</fpage><lpage>54</lpage><pub-id pub-id-type="doi">10.1007/978-981-10-3005-5_4</pub-id></nlm-citation></ref><ref id="ref29"><label>29</label><nlm-citation citation-type="web"><person-group person-group-type="author"><collab>Swin</collab></person-group><article-title>UNETR: swin transformers for semantic segmentation of brain tumors in MRI images</article-title><source>SpringerLink</source><year>2022</year><access-date>2024-08-08</access-date><comment><ext-link ext-link-type="uri" xlink:href="https://link.springer.com/chapter/10.1007/978-3-031-08999-2_22">https://link.springer.com/chapter/10.1007/978-3-031-08999-2_22</ext-link></comment></nlm-citation></ref><ref id="ref30"><label>30</label><nlm-citation citation-type="confproc"><person-group person-group-type="author"><name name-style="western"><surname>Li</surname><given-names>W</given-names> </name><name name-style="western"><surname>Yuille</surname><given-names>A</given-names> </name><name name-style="western"><surname>Zhou</surname><given-names>Z</given-names> </name></person-group><article-title>How well do supervised 3D models transfer to medical imaging tasks</article-title><conf-name>The Twelfth International Conference on Learning Representations</conf-name><conf-date>May 7-11, 2024</conf-date><conf-loc>Vienna, Austria</conf-loc><comment><ext-link ext-link-type="uri" xlink:href="https://openreview.net/forum?id=AhizIPytk4">https://openreview.net/forum?id=AhizIPytk4</ext-link></comment></nlm-citation></ref><ref id="ref31"><label>31</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Ma</surname><given-names>J</given-names> </name><name name-style="western"><surname>Zhang</surname><given-names>Y</given-names> </name><name name-style="western"><surname>Gu</surname><given-names>S</given-names> </name><etal/></person-group><article-title>AbdomenCT-1K: Is abdominal organ segmentation a solved problem?</article-title><source>IEEE Trans Pattern Anal Mach Intell</source><year>2022</year><month>10</month><volume>44</volume><issue>10</issue><fpage>6695</fpage><lpage>6714</lpage><pub-id pub-id-type="doi">10.1109/TPAMI.2021.3100536</pub-id><pub-id pub-id-type="medline">34314356</pub-id></nlm-citation></ref><ref id="ref32"><label>32</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Li</surname><given-names>W</given-names> </name><name name-style="western"><surname>Qu</surname><given-names>C</given-names> </name><name name-style="western"><surname>Chen</surname><given-names>X</given-names> </name><etal/></person-group><article-title>AbdomenAtlas: a large-scale, detailed-annotated, &#x0026; multi-center dataset for efficient transfer learning and open algorithmic benchmarking</article-title><source>Med Image Anal</source><year>2024</year><month>10</month><volume>97</volume><fpage>103285</fpage><pub-id pub-id-type="doi">10.1016/j.media.2024.103285</pub-id></nlm-citation></ref><ref id="ref33"><label>33</label><nlm-citation citation-type="confproc"><person-group person-group-type="author"><name name-style="western"><surname>Qu</surname><given-names>C</given-names> </name><name name-style="western"><surname>Zhang</surname><given-names>T</given-names> </name><name name-style="western"><surname>Qiao</surname><given-names>H</given-names> </name><etal/></person-group><article-title>AbdomenAtlas-8K: annotating 8,000 ct volumes for multi-organ segmentation in three weeks</article-title><year>2023</year><conf-name>The 37th International Conference on Neural Information Processing Systems</conf-name><conf-date>Dec 10-16, 2023</conf-date><conf-loc>New Orleans, USA</conf-loc></nlm-citation></ref><ref id="ref34"><label>34</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>L&#x00F6;ffler</surname><given-names>MT</given-names> </name><name name-style="western"><surname>Sekuboyina</surname><given-names>A</given-names> </name><name name-style="western"><surname>Jacob</surname><given-names>A</given-names> </name><etal/></person-group><article-title>A vertebral segmentation dataset with fracture grading</article-title><source>Radiol Artif Intell</source><year>2020</year><month>07</month><volume>2</volume><issue>4</issue><fpage>e190138</fpage><pub-id pub-id-type="doi">10.1148/ryai.2020190138</pub-id><pub-id pub-id-type="medline">33937831</pub-id></nlm-citation></ref><ref id="ref35"><label>35</label><nlm-citation citation-type="preprint"><person-group person-group-type="author"><name name-style="western"><surname>D&#x2019;Antonoli</surname><given-names>TA</given-names> </name><name name-style="western"><surname>Berger</surname><given-names>LK</given-names> </name><name name-style="western"><surname>Indrakanti</surname><given-names>AK</given-names> </name><etal/></person-group><article-title>TotalSegmentator MRI: sequence-independent segmentation of 59 anatomical structures in MR images</article-title><source>arXiv</source><comment>Preprint posted online on  Feb 26, 2025</comment><pub-id pub-id-type="doi">10.48550/arXiv.2405.19492</pub-id></nlm-citation></ref><ref id="ref36"><label>36</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Macdonald</surname><given-names>JA</given-names> </name><name name-style="western"><surname>Zhu</surname><given-names>Z</given-names> </name><name name-style="western"><surname>Konkel</surname><given-names>B</given-names> </name><name name-style="western"><surname>Mazurowski</surname><given-names>MA</given-names> </name><name name-style="western"><surname>Wiggins</surname><given-names>WF</given-names> </name><name name-style="western"><surname>Bashir</surname><given-names>MR</given-names> </name></person-group><article-title>Duke liver dataset: a publicly available liver MRI dataset with liver segmentation masks and series labels</article-title><source>Radiol Artif Intell</source><year>2023</year><month>09</month><volume>5</volume><issue>5</issue><fpage>e220275</fpage><pub-id pub-id-type="doi">10.1148/ryai.220275</pub-id><pub-id pub-id-type="medline">37795141</pub-id></nlm-citation></ref><ref id="ref37"><label>37</label><nlm-citation citation-type="web"><article-title>Dataset with segmentations of 117 important anatomical structures in 1228 CT images</article-title><source>Zenodo</source><year>2023</year><access-date>2025-02-04</access-date><comment><ext-link ext-link-type="uri" xlink:href="https://zenodo.org/records/10047292">https://zenodo.org/records/10047292</ext-link></comment></nlm-citation></ref></ref-list><app-group><supplementary-material id="app1"><label>Multimedia Appendix 1</label><p>Graphs and boxplots displaying mean areas of organs at various levels and comparison of DSCs.</p><media xlink:href="ai_v4i1e72109_app1.docx" xlink:title="DOCX File, 271 KB"/></supplementary-material></app-group></back></article>