Cross-Modal Graph Learning for Perivascular Spaces Segmentation

Research output: Chapter in Book/Report/Conference proceedingConference contributionAcademicpeer-review

Abstract

Perivascular spaces (PVS), also known as Virchow-Robin spaces, are critical biomarkers for diagnosing cerebral small vessel disease (CSVD). Quantifying PVS visible in magnetic resonance imaging (MRI) is essential for understanding their relationship with various neurological disorders. Traditional methods for assessing PVS rely on visual scoring of MRI images, which is time-consuming, subjective, and unsuitable for large-scale studies. Additionally, due to their small size, scattered distribution, and complex morphology, PVS can easily be confused with neighboring structures, posing significant challenges for their accurate extraction. In this paper, we propose a novel graph interaction-enhanced model based on vision-language modeling (VLM) technology for accurate PVS extraction from MRI. Our approach leverages textual information to guide image feature extraction and employs a graph structure to enhance cross-modal interactions, facilitating the reasoning of relationships between different modalities. Furthermore, we introduce a cross-modal attention mechanism for global feature alignment and an attention-based dynamic fusion module to effectively integrate multi-modal information, improving the accuracy of PVS segmentation. Validated on an independent T1-weighted dataset, our model demonstrates superior performance in capturing both global and local information, addressing the limitations of traditional image-only models and providing a robust solution for PVS segmentation in complex clinical scenarios.
Original languageEnglish
Title of host publicationMedical Image Computing and Computer Assisted Intervention – MICCAI 2025
Subtitle of host publication28th International Conference, Daejeon, South Korea, September 23–27, 2025, Proceedings, Part IV
EditorsJames C. Gee, Daniel C. Alexander, Jaesung Hong, Juan Eugenio Iglesias, Carole H. Sudre, Archana Venkataraman, Polina Golland, Jong Hyo Kim, Jinah Park
Place of PublicationCham
PublisherSpringer
Pages111-120
Number of pages10
ISBN (Electronic)978-3-032-04965-0
ISBN (Print)978-3-032-04964-3
DOIs
Publication statusPublished - 19 Sept 2025
Event28th International Conference on Medical Image Computing and Computer Assisted Intervention - Daejeon, Korea, Republic of
Duration: 23 Sept 202527 Sept 2025
https://conferences.miccai.org/2025/en/

Publication series

NameLecture Notes in Computer Science (LNCS)
Volume15963
ISSN (Print)0302-9743
ISSN (Electronic)1611-3349

Conference

Conference28th International Conference on Medical Image Computing and Computer Assisted Intervention
Abbreviated titleMICCAI
Country/TerritoryKorea, Republic of
CityDaejeon
Period23/09/2527/09/25
Internet address

Funding

This work was supported in part by the National Science Foundation Program of China (62371442), and in part by the Zhejiang Provincial Natural Science Foundation (LQ23F010002, LZ23F010002, LR24F010002).

Keywords

  • Cross Attention
  • GCN
  • MRI
  • PVS
  • VLM

Fingerprint

Dive into the research topics of 'Cross-Modal Graph Learning for Perivascular Spaces Segmentation'. Together they form a unique fingerprint.

Cite this