TY - JOUR
T1 - Deep learning generalization for diabetic retinopathy staging from fundus images
AU - Men, Yevgeniy
AU - Fhima, Jonathan
AU - Celi, Leo Anthony
AU - Ribeiro, Lucas Zago
AU - Nakayama, Luis Filipe
AU - Behar, Joachim A.
N1 - Publisher Copyright:
Creative Commons Attribution license.
PY - 2025/1/22
Y1 - 2025/1/22
N2 - Objective. Diabetic retinopathy (DR) is a serious diabetes complication that can lead to vision loss, making timely identification crucial. Existing data-driven algorithms for DR staging from digital fundus images (DFIs) often struggle with generalization due to distribution shifts between training and target domains.Approach. To address this, DRStageNet, a deep learning model, was developed using six public and independent datasets with 91 984 DFIs from diverse demographics. Five pretrained self-supervised vision transformers (ViTs) were benchmarked, with the best further trained using a multi-source domain (MSD) fine-tuning strategy.Main results. DINOv2 showed a 27.4% improvement in L-Kappa versus other pretrained ViT. MSD fine-tuning improved performance in four of five target domains. The error analysis revealing 60% of errors due to incorrect labels, 77.5% of which were correctly classified by DRStageNet.Significance. We developed DRStageNet, a DL model for DR, designed to accurately stage the condition while addressing the challenge of generalizing performance across target domains. The model and explainability heatmaps are available atwww.aimlab-technion.com/lirot-ai.
AB - Objective. Diabetic retinopathy (DR) is a serious diabetes complication that can lead to vision loss, making timely identification crucial. Existing data-driven algorithms for DR staging from digital fundus images (DFIs) often struggle with generalization due to distribution shifts between training and target domains.Approach. To address this, DRStageNet, a deep learning model, was developed using six public and independent datasets with 91 984 DFIs from diverse demographics. Five pretrained self-supervised vision transformers (ViTs) were benchmarked, with the best further trained using a multi-source domain (MSD) fine-tuning strategy.Main results. DINOv2 showed a 27.4% improvement in L-Kappa versus other pretrained ViT. MSD fine-tuning improved performance in four of five target domains. The error analysis revealing 60% of errors due to incorrect labels, 77.5% of which were correctly classified by DRStageNet.Significance. We developed DRStageNet, a DL model for DR, designed to accurately stage the condition while addressing the challenge of generalizing performance across target domains. The model and explainability heatmaps are available atwww.aimlab-technion.com/lirot-ai.
KW - deep learning
KW - diabetic retinopathy
KW - fundus image
KW - self-supervised learning
KW - transformers
UR - http://www.scopus.com/inward/record.url?scp=85216606882&partnerID=8YFLogxK
U2 - 10.1088/1361-6579/ada86a
DO - 10.1088/1361-6579/ada86a
M3 - ???researchoutput.researchoutputtypes.contributiontojournal.article???
C2 - 39788077
AN - SCOPUS:85216606882
SN - 0967-3334
VL - 13
JO - Physiological Measurement
JF - Physiological Measurement
IS - 1
ER -