Please use this identifier to cite or link to this item: http://hdl.handle.net/10397/118725
PIRA download icon_1.1View/Download Full Text
DC FieldValueLanguage
dc.contributorSchool of Optometryen_US
dc.contributorResearch Centre for SHARP Visionen_US
dc.creatorChen, Ren_US
dc.creatorZhang, Wen_US
dc.creatorLiu, Ben_US
dc.creatorWu, Xen_US
dc.creatorChen, Xen_US
dc.creatorXu, Pen_US
dc.creatorLiu, Sen_US
dc.creatorHe, Men_US
dc.creatorShi, Den_US
dc.date.accessioned2026-05-14T05:44:01Z-
dc.date.available2026-05-14T05:44:01Z-
dc.identifier.urihttp://hdl.handle.net/10397/118725-
dc.language.isoenen_US
dc.publisherNature Publishing Groupen_US
dc.rightsOpen Access This article is licensed under a Creative Commons Attribution-NonCommercial-NoDerivatives 4.0 International License, which permits any non-commercial use, sharing, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if you modified the licensed material. You do not have permission under this licence to share adapted material derived from this article or parts of it. The images or other third party material in this article are included in the article’s Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article’s Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by-nc-nd/4.0/.en_US
dc.rights© The Author(s) 2026en_US
dc.rightsThe following publication Chen, R., Zhang, W., Liu, B. et al. Boosting foundation models for rare eye disease diagnosis via a multimodal text-to-image generative framework. npj Digit. Med. 9, 371 (2026) is available at https://doi.org/10.1038/s41746-026-02560-2.en_US
dc.titleBoosting foundation models for rare eye disease diagnosis via a multimodal text-to-image generative frameworken_US
dc.typeJournal/Magazine Articleen_US
dc.identifier.volume9en_US
dc.identifier.doi10.1038/s41746-026-02560-2en_US
dcterms.abstractThe rising prevalence of vision-threatening retinal diseases poses a significant burden on the global healthcare systems. Though deep learning (DL) techniques offer promising avenues for improving diagnostic efficiency, data scarcity and imbalance issues persist in training robust diagnostic models, particularly for rare eye diseases. Here, we introduce EyeDiff, a generative foundation model capable of synthesizing lesion-preserving ophthalmic images from textual descriptions. Both objective metrics and expert human evaluations confirmed EyeDiff’s ability to generate high-fidelity images across multiple imaging modalities, accurately reflecting textual descriptions of diverse retinal diseases and lesion types. By augmenting minority classes across 11 globally sourced datasets, EyeDiff consistently boosted the diagnostic accuracy for both common and rare eye diseases across different foundation model types, including modality-specific, multimodal and vision-language foundation models trained solely on real data. These results underscore EyeDiff’s potential as a general-purpose text-to-image foundation model, offering a scalable and flexible approach to generate balanced, disease-relevant data for advancing retinal disease diagnosis.en_US
dcterms.accessRightsopen accessen_US
dcterms.bibliographicCitationnpj digital medicine, 2026, v. 9, 371en_US
dcterms.isPartOfnpj digital medicineen_US
dcterms.issued2026-
dc.identifier.eissn2398-6352en_US
dc.identifier.artn371en_US
dc.description.validate202605 bcchen_US
dc.description.oaVersion of Recorden_US
dc.identifier.FolderNumbera4418-
dc.identifier.SubFormID52749-
dc.description.fundingSourceOthersen_US
dc.description.fundingTextWe thank the American Society of Retina Specialists for providing the valuable Retina Image Bank and the InnoHK HKSAR Government for providing valuable support. The study was supported by the Start-up Fund for RAPs under the Strategic Hiring Scheme (P0048623) from HKSAR, Global STEM Professorship Scheme (P0046113), and Henry G. Leong Endowed Professorship in Elderly Vision Health. The sponsors or funding organizations had no role in the design or conduct of this research.en_US
dc.description.pubStatusPublisheden_US
dc.description.oaCategoryCCen_US
dc.relation.rdataThe data for model training in the current study are available as open data through the following links: Retinal Image Bank (https://imagebank.asrs.org/), EyePACS (https://www.kaggle.com/c/diabetic-retinopathy-detection/data), OCTDL (https://ieee-dataport.org/documents/octdl-optical-coherence-tomography-dataset-image-based-deep-learning-methods), REFUGE (https://bitbucket.org/woalsdnd/refuge/src/master/), ORIGA (https://figshare.com/articles/dataset/Retinal_Fundus_Glaucoma_Image_dataset/24549217en_US
Appears in Collections:Journal/Magazine Article
Files in This Item:
File Description SizeFormat 
s41746-026-02560-2.pdf4.33 MBAdobe PDFView/Open
Open Access Information
Status open access
File Version Version of Record
Access
View full-text via PolyU eLinks SFX Query
Show simple item record

Google ScholarTM

Check

Altmetric


Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.