Uncertainty-Aware Adapter: Adapting Segment Anything Model (SAM) for Ambiguous Medical Image Segmentation
The Segment Anything Model (SAM) gained significant success in natural image segmentation, and many methods have tried to fine-tune it to medical image segmentation. An efficient way to do so is by using Adapters, specialized modules that learn just a few parameters to tailor SAM specifically for me...
Saved in:
Main Authors | , , , , , |
---|---|
Format | Journal Article |
Language | English |
Published |
16.03.2024
|
Subjects | |
Online Access | Get full text |
Cover
Loading…
Summary: | The Segment Anything Model (SAM) gained significant success in natural image
segmentation, and many methods have tried to fine-tune it to medical image
segmentation. An efficient way to do so is by using Adapters, specialized
modules that learn just a few parameters to tailor SAM specifically for medical
images. However, unlike natural images, many tissues and lesions in medical
images have blurry boundaries and may be ambiguous. Previous efforts to adapt
SAM ignore this challenge and can only predict distinct segmentation. It may
mislead clinicians or cause misdiagnosis, especially when encountering rare
variants or situations with low model confidence. In this work, we propose a
novel module called the Uncertainty-aware Adapter, which efficiently
fine-tuning SAM for uncertainty-aware medical image segmentation. Utilizing a
conditional variational autoencoder, we encoded stochastic samples to
effectively represent the inherent uncertainty in medical imaging. We designed
a new module on a standard adapter that utilizes a condition-based strategy to
interact with samples to help SAM integrate uncertainty. We evaluated our
method on two multi-annotated datasets with different modalities: LIDC-IDRI
(lung abnormalities segmentation) and REFUGE2 (optic-cup segmentation). The
experimental results show that the proposed model outperforms all the previous
methods and achieves the new state-of-the-art (SOTA) on both benchmarks. We
also demonstrated that our method can generate diverse segmentation hypotheses
that are more realistic as well as heterogeneous. |
---|---|
DOI: | 10.48550/arxiv.2403.10931 |