Please use this identifier to cite or link to this item: http://hdl.handle.net/10397/117678
PIRA download icon_1.1View/Download Full Text
DC FieldValueLanguage
dc.contributorSchool of Optometryen_US
dc.contributorDepartment of Computingen_US
dc.contributorResearch Centre for SHARP Visionen_US
dc.creatorWu, Xen_US
dc.creatorWang, Len_US
dc.creatorChen, Ren_US
dc.creatorLiu, Ben_US
dc.creatorZhang, Wen_US
dc.creatorYang, Xen_US
dc.creatorFeng, Yen_US
dc.creatorHe, Men_US
dc.creatorShi, Den_US
dc.date.accessioned2026-02-26T04:26:50Z-
dc.date.available2026-02-26T04:26:50Z-
dc.identifier.issn2168-6165en_US
dc.identifier.urihttp://hdl.handle.net/10397/117678-
dc.language.isoenen_US
dc.publisherAmerican Medical Associationen_US
dc.rights©2025 American Medical Association. All rights reserved, including those for text and data mining, AI training, and similar technologies.en_US
dc.rightsThe following publication Wu X, Wang L, Chen R, et al. Generation of Fundus Fluorescein Angiography Videos for Health Care Data Sharing. JAMA Ophthalmol. 2025;143(8):623–632 is available at https://doi.org/10.1001/jamaophthalmol.2025.1419.en_US
dc.titleGeneration of fundus fluorescein angiography videos for health care data sharingen_US
dc.typeJournal/Magazine Articleen_US
dc.description.otherinformationTitle on author's file: FFA Sora: generating fundus fluorescein angiography videos for healthcare data sharingen_US
dc.identifier.spage623en_US
dc.identifier.epage632en_US
dc.identifier.volume143en_US
dc.identifier.issue8en_US
dc.identifier.doi10.1001/jamaophthalmol.2025.1419en_US
dcterms.abstractImportance: Medical data sharing faces strict restrictions. Text-to-video generation shows potential for creating realistic medical data while preserving privacy, offering a solution for cross-center data sharing and medical education.en_US
dcterms.abstractObjective: To develop and evaluate a text-to-video generative artificial intelligence (AI)–driven model that converts the text of reports into dynamic fundus fluorescein angiography (FFA) videos, enabling visualization of retinal vascular and structural abnormalities.en_US
dcterms.abstractDesign, Setting, and Participants: This study retrospectively collected anonymized FFA data from a tertiary hospital in China. The dataset included both the medical records and FFA examinations of patients assessed between November 2016 and December 2019. A text-to-video model was developed and evaluated. The AI-driven model integrated the wavelet-flow variational autoencoder and the diffusion transformer.en_US
dcterms.abstractMain Outcomes and Measures: The AI-driven model’s performance was assessed through objective metrics (Fréchet video distance, learned perceptual image patch similarity score, and visual question answering score [VQAScore]). The domain-specific evaluation for the generated FFA videos was measured by the bidirectional encoder representations from transformers score (BERTScore). Image retrieval was evaluated using a Recall@K score. Each video was rated for quality by 3 ophthalmologists on a scale of 1 (excellent) to 5 (very poor).en_US
dcterms.abstractResults: A total of 3625 FFA videos were included (2851 videos [78.6%] for training, 387 videos [10.7%] for validation, and 387 videos [10.7%] for testing). The AI-generated FFA videos demonstrated retinal abnormalities from the input text (Fréchet video distance of 2273, a mean learned perceptual image patch similarity score of 0.48 [SD, 0.04], and a mean VQAScore of 0.61 [SD, 0.08]). The domain-specific evaluations showed alignment between the generated videos and textual prompts (mean BERTScore, 0.35 [SD, 0.09]). The Recall@K scores were 0.02 for K = 5, 0.04 for K = 10, and 0.16 for K = 50, yielding a mean score of 0.073, reflecting disparities between AI-generated and real clinical videos and demonstrating privacy-preserving effectiveness. For assessment of visual quality of the FFA videos by the 3 ophthalmologists, the mean score was 1.57 (SD, 0.44).en_US
dcterms.abstractConclusions and Relevance: This study demonstrated that an AI-driven text-to-video model generated FFA videos from textual descriptions, potentially improving visualization for clinical and educational purposes. The privacy-preserving nature of the model may address key challenges in data sharing while trying to ensure compliance with confidentiality standards.en_US
dcterms.accessRightsopen accessen_US
dcterms.bibliographicCitationJAMA ophthalmology, Aug. 2025, v. 143, no. 8, p. 623-632en_US
dcterms.isPartOfJAMA ophthalmologyen_US
dcterms.issued2025-08-
dc.identifier.scopus2-s2.0-105010101144-
dc.identifier.pmid40569610-
dc.identifier.eissn2168-6173en_US
dc.description.validate202602 bcchen_US
dc.description.oaAccepted Manuscripten_US
dc.identifier.SubFormIDG001080/2026-02-
dc.description.fundingSourceOthersen_US
dc.description.fundingTextDr He received research support from the Global STEM Professorship Scheme (P0046113) and the Henry G. Leong Endowed Professorship in Elderly Vision Health. This work was also funded by the Hong Kong Jockey Club Charities Trust.en_US
dc.description.pubStatusPublisheden_US
dc.description.oaCategoryGreen (AAM)en_US
Appears in Collections:Journal/Magazine Article
Files in This Item:
File Description SizeFormat 
Wu_Generation_Fundus_Fluorescein_.pdfPre-Published version4.17 MBAdobe PDFView/Open
Open Access Information
Status open access
File Version Final Accepted Manuscript
Access
View full-text via PolyU eLinks SFX Query
Show simple item record

SCOPUSTM   
Citations

4
Citations as of May 8, 2026

WEB OF SCIENCETM
Citations

5
Citations as of Apr 23, 2026

Google ScholarTM

Check

Altmetric


Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.