Please use this identifier to cite or link to this item: http://hdl.handle.net/10397/114034
PIRA download icon_1.1View/Download Full Text
DC FieldValueLanguage
dc.contributorSchool of Optometryen_US
dc.contributorResearch Centre for SHARP Visionen_US
dc.creatorShi, Den_US
dc.creatorZhang, Wen_US
dc.creatorYang, Jen_US
dc.creatorHuang, Sen_US
dc.creatorChen, Xen_US
dc.creatorXu, Pen_US
dc.creatorJin, Ken_US
dc.creatorLin, Sen_US
dc.creatorWei, Jen_US
dc.creatorYusufu, Men_US
dc.creatorLiu, Sen_US
dc.creatorZhang, Qen_US
dc.creatorGe, Zen_US
dc.creatorXu, Xen_US
dc.creatorHe, Men_US
dc.date.accessioned2025-07-10T03:59:38Z-
dc.date.available2025-07-10T03:59:38Z-
dc.identifier.urihttp://hdl.handle.net/10397/114034-
dc.language.isoenen_US
dc.publisherNature Publishing Groupen_US
dc.rights© The Author(s) 2025en_US
dc.rightsThis article is licensed under a Creative Commons Attribution-NonCommercial-NoDerivatives 4.0 International License, which permits any non-commercial use, sharing, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if you modified the licensed material. You do not have permission under this licence to share adapted material derived from this article or parts of it. The images or other third party material in this article are included in the article’s Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article’s Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by-nc-nd/4.0/.en_US
dc.rightsThe following publication Shi, D., Zhang, W., Yang, J. et al. A multimodal visual–language foundation model for computational ophthalmology. npj Digit. Med. 8, 381 (2025) is available at https://doi.org/10.1038/s41746-025-01772-2.en_US
dc.titleA multimodal visual–language foundation model for computational ophthalmologyen_US
dc.typeJournal/Magazine Articleen_US
dc.identifier.volume8en_US
dc.identifier.doi10.1038/s41746-025-01772-2en_US
dcterms.abstractEarly detection of eye diseases is vital for preventing vision loss. Existing ophthalmic artificial intelligence models focus on single modalities, overlooking multi-view information and struggling with rare diseases due to long-tail distributions. We propose EyeCLIP, a multimodal visual-language foundation model trained on 2.77 million ophthalmology images from 11 modalities with partial clinical text. Our novel pretraining strategy combines self-supervised reconstruction, multimodal image contrastive learning, and image-text contrastive learning to capture shared representations across modalities. EyeCLIP demonstrates robust performance across 14 benchmark datasets, excelling in disease classification, visual question answering, and cross-modal retrieval. It also exhibits strong few-shot and zero-shot capabilities, enabling accurate predictions in real-world, long-tail scenarios. EyeCLIP offers significant potential for detecting both ocular and systemic diseases, and bridging gaps in real-world clinical applications.en_US
dcterms.accessRightsopen accessen_US
dcterms.bibliographicCitationnpj digital medicine, 2025, v. 8, 381en_US
dcterms.isPartOfnpj digital medicineen_US
dcterms.issued2025-
dc.identifier.scopus2-s2.0-105008508876-
dc.identifier.eissn2398-6352en_US
dc.identifier.artn381en_US
dc.description.validate202507 bcwhen_US
dc.description.oaVersion of Recorden_US
dc.identifier.FolderNumbera3874, a3849e-
dc.identifier.SubFormID51486, 51398-
dc.description.fundingSourceSelf-fundeden_US
dc.description.pubStatusPublisheden_US
dc.description.oaCategoryCCen_US
Appears in Collections:Journal/Magazine Article
Files in This Item:
File Description SizeFormat 
s41746-025-01772-2.pdf3.1 MBAdobe PDFView/Open
Open Access Information
Status open access
File Version Version of Record
Access
View full-text via PolyU eLinks SFX Query
Show simple item record

SCOPUSTM   
Citations

6
Citations as of Oct 10, 2025

Google ScholarTM

Check

Altmetric


Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.