Cross-Subject Reference Attention for Brain Lesion Segmentation
Brain lesion segmentation serves as a useful tool for clinical diagnosis and scientific research. The Transformer architecture has achieved remarkable performance in the field of computer vision and can potentially benefit brain lesion segmentation. Standard Transformer models partition an image int...
Saved in:
Published in | 2025 IEEE 2nd International Conference on Deep Learning and Computer Vision (DLCV) pp. 1 - 5 |
---|---|
Main Authors | , |
Format | Conference Proceeding |
Language | English |
Published |
IEEE
06.06.2025
|
Subjects | |
Online Access | Get full text |
DOI | 10.1109/DLCV65218.2025.11088861 |
Cover
Loading…
Abstract | Brain lesion segmentation serves as a useful tool for clinical diagnosis and scientific research. The Transformer architecture has achieved remarkable performance in the field of computer vision and can potentially benefit brain lesion segmentation. Standard Transformer models partition an image into patches and compute the self-attention matrix between these patches to model long-range dependencies. However, as brain lesions usually occupy only a small portion of the image, self-attention is computed mostly between patches of lesions and healthy tissue. Consequently, there is a lack of long-range dependency that is relevant to brain lesion segmentation, and existing Transformer models have not achieved a clear improvement over convolution-based methods, such as nnU-Net. To address the limitation above, we propose a cross-subject reference attention (CRA) mechanism for brain lesion segmentation. CRA exploits the information from not only the image to segment but also reference lesion information from other images, which can provide relevant long-range dependency. Specifically, CRA consists of a subject-level tokenizer, a reference pool, and a hierarchical cross-attention module. The subject-level tokenizer first maps the input feature map of each subject into a fixed number of tokens, which compresses the rich image information to reduce the subsequent computation overhead. Then, the tokens of selected subjects are stored in the reference pool as reference tokens. Finally, in the hierarchical cross-attention image features are adjusted with adaptive guidance of reference tokens, which alleviates the lack of lesion patches in a single image. CRA is agnostic to the segmentation backbone and we integrate it with the state-of-the-art nnU-Net framework. To evaluate the proposed method, we performed experiments on multiple datasets, and the results indicate that CRA leads to improved accuracy of brain lesion segmentation. |
---|---|
AbstractList | Brain lesion segmentation serves as a useful tool for clinical diagnosis and scientific research. The Transformer architecture has achieved remarkable performance in the field of computer vision and can potentially benefit brain lesion segmentation. Standard Transformer models partition an image into patches and compute the self-attention matrix between these patches to model long-range dependencies. However, as brain lesions usually occupy only a small portion of the image, self-attention is computed mostly between patches of lesions and healthy tissue. Consequently, there is a lack of long-range dependency that is relevant to brain lesion segmentation, and existing Transformer models have not achieved a clear improvement over convolution-based methods, such as nnU-Net. To address the limitation above, we propose a cross-subject reference attention (CRA) mechanism for brain lesion segmentation. CRA exploits the information from not only the image to segment but also reference lesion information from other images, which can provide relevant long-range dependency. Specifically, CRA consists of a subject-level tokenizer, a reference pool, and a hierarchical cross-attention module. The subject-level tokenizer first maps the input feature map of each subject into a fixed number of tokens, which compresses the rich image information to reduce the subsequent computation overhead. Then, the tokens of selected subjects are stored in the reference pool as reference tokens. Finally, in the hierarchical cross-attention image features are adjusted with adaptive guidance of reference tokens, which alleviates the lack of lesion patches in a single image. CRA is agnostic to the segmentation backbone and we integrate it with the state-of-the-art nnU-Net framework. To evaluate the proposed method, we performed experiments on multiple datasets, and the results indicate that CRA leads to improved accuracy of brain lesion segmentation. |
Author | Jiang, Runze Ye, Chuyang |
Author_xml | – sequence: 1 givenname: Runze surname: Jiang fullname: Jiang, Runze organization: School of Integrated Circuits and Electronics, Beijing Institute of Technology,Beijing,China – sequence: 2 givenname: Chuyang surname: Ye fullname: Ye, Chuyang organization: School of Integrated Circuits and Electronics, Beijing Institute of Technology,Beijing,China |
BookMark | eNo1j91Kw0AQhVfQC237BoJ5gdSdGTc7eyU1WhUCBVu8LfmZlRW7kU288O1tUK8OfN_hwLlQp7GPotQV6CWAdtf3VflaGAReokYzMWYu4EQtnHVMBAaxcHyubsvUD0O-_WrepR2zF_GSJLaSrcZR4hj6mPk-ZXepDjGrZJjAVt4OR1dPdq7OfP0xyOIvZ2q3ftiVT3m1eXwuV1UeHI05gyeLBWmAjqzV4moEw56wNebGaM-mM02HLSF3cuywY9TgwRfcWHQ0U5e_s0FE9p8pHOr0vf-_RT94lkYo |
ContentType | Conference Proceeding |
DBID | 6IE 6IL CBEJK RIE RIL |
DOI | 10.1109/DLCV65218.2025.11088861 |
DatabaseName | IEEE Electronic Library (IEL) Conference Proceedings IEEE Proceedings Order Plan All Online (POP All Online) 1998-present by volume IEEE Xplore All Conference Proceedings IEEE Electronic Library (IEL) IEEE Proceedings Order Plans (POP All) 1998-Present |
DatabaseTitleList | |
Database_xml | – sequence: 1 dbid: RIE name: IEEE Electronic Library (IEL) url: https://proxy.k.utb.cz/login?url=https://ieeexplore.ieee.org/ sourceTypes: Publisher |
DeliveryMethod | fulltext_linktorsrc |
EISBN | 9798331522698 |
EndPage | 5 |
ExternalDocumentID | 11088861 |
Genre | orig-research |
GroupedDBID | 6IE 6IL CBEJK RIE RIL |
ID | FETCH-LOGICAL-i93t-81f37263011d3770e9a2158f32c55450f85d5bd2c328de1d3898201f1f68b7293 |
IEDL.DBID | RIE |
IngestDate | Wed Aug 20 06:20:57 EDT 2025 |
IsPeerReviewed | false |
IsScholarly | false |
Language | English |
LinkModel | DirectLink |
MergedId | FETCHMERGED-LOGICAL-i93t-81f37263011d3770e9a2158f32c55450f85d5bd2c328de1d3898201f1f68b7293 |
PageCount | 5 |
ParticipantIDs | ieee_primary_11088861 |
PublicationCentury | 2000 |
PublicationDate | 2025-June-6 |
PublicationDateYYYYMMDD | 2025-06-06 |
PublicationDate_xml | – month: 06 year: 2025 text: 2025-June-6 day: 06 |
PublicationDecade | 2020 |
PublicationTitle | 2025 IEEE 2nd International Conference on Deep Learning and Computer Vision (DLCV) |
PublicationTitleAbbrev | DLCV |
PublicationYear | 2025 |
Publisher | IEEE |
Publisher_xml | – name: IEEE |
Score | 1.9136558 |
Snippet | Brain lesion segmentation serves as a useful tool for clinical diagnosis and scientific research. The Transformer architecture has achieved remarkable... |
SourceID | ieee |
SourceType | Publisher |
StartPage | 1 |
SubjectTerms | Accuracy Attention Brain lesion segmentation Brain modeling Computational modeling Computer vision Cross-subject reference Image coding Image segmentation Lesions Navigation Transformers |
Title | Cross-Subject Reference Attention for Brain Lesion Segmentation |
URI | https://ieeexplore.ieee.org/document/11088861 |
hasFullText | 1 |
inHoldings | 1 |
isFullTextHit | |
isPrint | |
link | http://utb.summon.serialssolutions.com/2.0.0/link/0/eLvHCXMwjV3LSgMxFA3qypWKFd9k4TbTTN5ZiVZLkVIEq3RXJi8RcSoy3fj1JplpRUFwF0IgCXmcm5tz7gXgwgrNbPAYBVwRxBzjqOKUIUMq7LismK4y22IiRo_sbsZnnVg9a2G895l85otUzH_5bmGXyVXWT5R1pdJjZzPus1as1XG2Sqz7N-PBk4hwlBhbhBer1j_ypmTYGO6AyarDli3yWiwbU9jPX7EY_z2iXdD7VujB-zX27IENX--Dy0GCPBTvguRcgesYsvCqaVpWI4wmKrxOWSHg2Cc_GXzwz2-d_KjugenwdjoYoS5BAnrRtEGqDFQSkY6oo1Jir6sI4CpQYqORwHFQ3HHjiKVEOR_bKJ3wPpRBKBONanoAtupF7Q8BpFRQWjorpbWMOKwdCaU3UjiuifHsCPTS5OfvbQiM-Wrex3_Un4DttAaZUyVOwVbzsfRnEb0bc55X7Qt6l5nT |
linkProvider | IEEE |
linkToHtml | http://utb.summon.serialssolutions.com/2.0.0/link/0/eLvHCXMwjV1LSwMxEA5SD3pSseLbHLxmu5vXJifRaqm6FsEqvZXNS0Tcimwv_nqT7LaiIHgLIZAMIfkmk--bAeBUc0m1sylyaYkRNZShkhGKFC5Tw_KSyjKyLUZ8-EhvJmzSitWjFsZaG8lnNgnN-JdvZnoeQmW9QFkXIjx2Vj3wU9bItVrWVpbK3mXRf-IekAJnC7NkMf5H5ZQIHIMNMFpM2fBFXpN5rRL9-Ssb47_XtAm63xo9eL9Eny2wYqttcNYPoIf8bRDCK3CZRRae13XDa4TeSYUXoS4ELGyIlMEH-_zWCpCqLhgPrsb9IWpLJKAXSWokMkdyzMMhNSTPUytLD-HCEay9m8BSJ5hhymBNsDDWjxEyIL7LHBfKu9VkB3SqWWV3ASSEE5IZnedaU2xSabDLrMq5YRIrS_dANxg_fW-SYEwXdu__0X8C1obju2JaXI9uD8B62I_IsOKHoFN_zO2Rx_JaHccd_AK8Rp0g |
openUrl | ctx_ver=Z39.88-2004&ctx_enc=info%3Aofi%2Fenc%3AUTF-8&rfr_id=info%3Asid%2Fsummon.serialssolutions.com&rft_val_fmt=info%3Aofi%2Ffmt%3Akev%3Amtx%3Abook&rft.genre=proceeding&rft.title=2025+IEEE+2nd+International+Conference+on+Deep+Learning+and+Computer+Vision+%28DLCV%29&rft.atitle=Cross-Subject+Reference+Attention+for+Brain+Lesion+Segmentation&rft.au=Jiang%2C+Runze&rft.au=Ye%2C+Chuyang&rft.date=2025-06-06&rft.pub=IEEE&rft.spage=1&rft.epage=5&rft_id=info:doi/10.1109%2FDLCV65218.2025.11088861&rft.externalDocID=11088861 |