Skip to main navigation Skip to search Skip to main content

Multimodality for Diagnosis of Asian Choroidal Vasculopathy: Results from a Novel Dataset and Deep-Learning Experiments

Research output: Chapter in Book/Report/Conference proceedingConference contribution

Abstract

Deep learning algorithms show tremendous potential for clinical decision-making - for example, in providing automated diagnoses of imaging data. However, typical clinical datasets often are limited in size, modalities, and contain heterogeneous, incomplete data, which presents challenges for deep learning frameworks that necessitate larger, uniform datasets, complicating their deployment especially with new types of disease models. In this work, we present a case study for deep learning in such a challenging setting in the context of diagnosing Asian choroidal Vasculopathy (ACV), which is a retinopathy profile currently under discussion in ophthalmology to be differentiated from age-related macular degeneration (AMD). We first introduce a novel, human-annotated multimodal dataset for ACV versus AMD diagnosis incorporating four different imaging modalities. We next explore the usefulness of “foundation models” for this data, compared to traditional dataset-specific training. Most importantly, we investigate which of the four modalities is most discriminative and whether bi-modal classification is able to enhance performance across multiple fusion approaches. We also discuss first results of salient features using explainability techniques.

Original languageEnglish
Title of host publicationMedical Image Computing and Computer Assisted Intervention – MICCAI 2024 Workshops - LDTM 2024, MMMI/ML4MHD 2024, ML-CDS 2024, Held in Conjunction with MICCAI 2024, Proceedings
EditorsAnna Schroder, Xiang Li, Tanveer Syeda-Mahmood, Neil P. Oxtoby, Alexandra Young, Alessa Hering, Tejas S. Mathai, Pritam Mukherjee, Sven Kuckertz, Tiantian He, Isaac Llorente-Saguer, Andreas Maier, Satyananda Kashyap, Hayit Greenspan, Anant Madabhushi
PublisherSpringer Science and Business Media Deutschland GmbH
Pages235-247
Number of pages13
ISBN (Print)9783031845246
DOIs
Publication statusPublished - 2025
EventWorkshop on Longitudinal Disease Tracking and Modeling with Medical Images and Data, LDTM 2024, 5th International Workshop on Multiscale Multimodal Medical Imaging, MMMI 2024, 1st Workshop on Machine Learning for Multimodal/-sensor Healthcare Data, ML4MHD2024 and Workshop on Multimodal Learning and Fusion Across Scales for Clinical Decision Support, ML-CDS 2024 held in conjunction with the 27th International Conference on Medical Image Computing and Computer Assisted Intervention, MICCAI 2024 - Marrakesh, Morocco
Duration: 2024 Oct 62024 Oct 10

Publication series

NameLecture Notes in Computer Science
Volume15401 LNCS
ISSN (Print)0302-9743
ISSN (Electronic)1611-3349

Conference

ConferenceWorkshop on Longitudinal Disease Tracking and Modeling with Medical Images and Data, LDTM 2024, 5th International Workshop on Multiscale Multimodal Medical Imaging, MMMI 2024, 1st Workshop on Machine Learning for Multimodal/-sensor Healthcare Data, ML4MHD2024 and Workshop on Multimodal Learning and Fusion Across Scales for Clinical Decision Support, ML-CDS 2024 held in conjunction with the 27th International Conference on Medical Image Computing and Computer Assisted Intervention, MICCAI 2024
Country/TerritoryMorocco
CityMarrakesh
Period24/10/624/10/10

Bibliographical note

Publisher Copyright:
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2025.

UN SDGs

This output contributes to the following UN Sustainable Development Goals (SDGs)

  1. SDG 3 - Good Health and Well-being
    SDG 3 Good Health and Well-being

Keywords

  • Deep Learning
  • Multimodal Imaging
  • Retinopathy

ASJC Scopus subject areas

  • Theoretical Computer Science
  • General Computer Science

Fingerprint

Dive into the research topics of 'Multimodality for Diagnosis of Asian Choroidal Vasculopathy: Results from a Novel Dataset and Deep-Learning Experiments'. Together they form a unique fingerprint.

Cite this