Abstract:Self-supervised learning (SSL) has enabled Vision Transformers (ViTs) to learn robust representations from large-scale natural image datasets, enhancing their generalization across domains. In retinal imaging, foundation models pretrained on either natural or ophthalmic data have shown promise, but the benefits of in-domain pretraining remain uncertain. To investigate this, we benchmark six SSL-pretrained ViTs on seven digital fundus image (DFI) datasets totaling 70,000 expert-annotated images for the task of moderate-to-late age-related macular degeneration (AMD) identification. Our results show that iBOT pretrained on natural images achieves the highest out-of-distribution generalization, with AUROCs of 0.80-0.97, outperforming domain-specific models, which achieved AUROCs of 0.78-0.96 and a baseline ViT-L with no pretraining, which achieved AUROCs of 0.68-0.91. These findings highlight the value of foundation models in improving AMD identification and challenge the assumption that in-domain pretraining is necessary. Furthermore, we release BRAMD, an open-access dataset (n=587) of DFIs with AMD labels from Brazil.
Abstract:The Hillel Yaffe Age Related Macular Degeneration (HYAMD) dataset is a longitudinal collection of 1,560 Digital Fundus Images (DFIs) from 325 patients examined at the Hillel Yaffe Medical Center (Hadera, Israel) between 2021 and 2024. The dataset includes an AMD cohort of 147 patients (aged 54-94) with varying stages of AMD and a control group of 190 diabetic retinopathy (DR) patients (aged 24-92). AMD diagnoses were based on comprehensive clinical ophthalmic evaluations, supported by Optical Coherence Tomography (OCT) and OCT angiography. Non-AMD DFIs were sourced from DR patients without concurrent AMD, diagnosed using macular OCT, fluorescein angiography, and widefield imaging. HYAMD provides gold-standard annotations, ensuring AMD labels were assigned following a full clinical assessment. Images were captured with a DRI OCT Triton (Topcon) camera, offering a 45 deg field of view and 1960 x 1934 pixel resolution. To the best of our knowledge, HYAMD is the first open-access retinal dataset from an Israeli sample, designed to support AMD identification using machine learning models.
Abstract:Glaucomatous optic neuropathy (GON) is a prevalent ocular disease that can lead to irreversible vision loss if not detected early and treated. The traditional diagnostic approach for GON involves a set of ophthalmic examinations, which are time-consuming and require a visit to an ophthalmologist. Recent deep learning models for automating GON detection from digital fundus images (DFI) have shown promise but often suffer from limited generalizability across different ethnicities, disease groups and examination settings. To address these limitations, we introduce GONet, a robust deep learning model developed using seven independent datasets, including over 119,000 DFIs with gold-standard annotations and from patients of diverse geographic backgrounds. GONet consists of a DINOv2 pre-trained self-supervised vision transformers fine-tuned using a multisource domain strategy. GONet demonstrated high out-of-distribution generalizability, with an AUC of 0.85-0.99 in target domains. GONet performance was similar or superior to state-of-the-art works and was significantly superior to the cup-to-disc ratio, by up to 21.6%. GONet is available at [URL provided on publication]. We also contribute a new dataset consisting of 768 DFI with GON labels as open access.