liu.seSearch for publications in DiVA
Change search
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • oxford
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf
LANGUAGE GUIDED DOMAIN GENERALIZED MEDICAL IMAGE SEGMENTATION
Mohamed Bin Zayed Univ Artif Intelligence, U Arab Emirates.
Mohamed Bin Zayed Univ Artif Intelligence, U Arab Emirates.
Mohamed Bin Zayed Univ Artif Intelligence, U Arab Emirates.
Linköping University, Department of Electrical Engineering, Computer Vision. Linköping University, Faculty of Science & Engineering. Mohamed Bin Zayed Univ Artif Intelligence, U Arab Emirates.
2024 (English)In: IEEE INTERNATIONAL SYMPOSIUM ON BIOMEDICAL IMAGING, ISBI 2024, IEEE , 2024Conference paper, Published paper (Refereed)
Abstract [en]

Single source domain generalization (SDG) holds promise for more reliable and consistent image segmentation across real-world clinical settings particularly in the medical domain, where data privacy and acquisition cost constraints often limit the availability of diverse datasets. Depending solely on visual features hampers the model's capacity to adapt effectively to various domains, primarily because of the presence of spurious correlations and domain-specific characteristics embedded within the image features. Incorporating text features alongside visual features is a potential solution to enhance the model's understanding of the data, as it goes beyond pixel-level information to provide valuable context. Textual cues describing the anatomical structures, their appearances, and variations across various imaging modalities can guide the model in domain adaptation, ultimately contributing to more robust and consistent segmentation. In this paper, we propose an approach that explicitly leverages textual information by incorporating a contrastive learning mechanism guided by the text encoder features to learn a more robust feature representation. We assess the effectiveness of our text-guided contrastive feature alignment technique in various scenarios, including cross-modality, cross-sequence, and cross-site settings for different segmentation tasks. Our approach achieves favorable performance against existing methods in literature. Our code and models will be publicly released.

Place, publisher, year, edition, pages
IEEE , 2024.
Series
IEEE International Symposium on Biomedical Imaging, ISSN 1945-7928, E-ISSN 1945-8452
Keywords [en]
Multi-modal contrastive learning; Medical image segmentation; Single source domain generalization
National Category
Medical Imaging
Identifiers
URN: urn:nbn:se:liu:diva-212430DOI: 10.1109/ISBI56570.2024.10635823ISI: 001305705103123Scopus ID: 2-s2.0-85203316739ISBN: 9798350313345 (print)ISBN: 9798350313338 (electronic)OAI: oai:DiVA.org:liu-212430DiVA, id: diva2:1945909
Conference
21st IEEE International Symposium on Biomedical Imaging (ISBI), Athens, GREECE, may 27-30, 2024
Note

Funding Agencies|Mohammed bin Zayed University of Articial Intelligence

Available from: 2025-03-19 Created: 2025-03-19 Last updated: 2025-03-19

Open Access in DiVA

No full text in DiVA

Other links

Publisher's full textScopus

Search in DiVA

By author/editor
Khan, Fahad
By organisation
Computer VisionFaculty of Science & Engineering
Medical Imaging

Search outside of DiVA

GoogleGoogle Scholar

doi
isbn
urn-nbn

Altmetric score

doi
isbn
urn-nbn
Total: 40 hits
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • oxford
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf