Please use this identifier to cite or link to this item: http://hdl.handle.net/10397/109379
DC FieldValueLanguage
dc.contributorDepartment of Biomedical Engineeringen_US
dc.contributorResearch Institute for Smart Ageingen_US
dc.creatorLi, Qen_US
dc.creatorYan, Xen_US
dc.creatorXu, Jen_US
dc.creatorYuan, Ren_US
dc.creatorZhang, Yen_US
dc.creatorFeng, Ren_US
dc.creatorShen, Qen_US
dc.creatorZhang, Xen_US
dc.creatorWang, Sen_US
dc.date.accessioned2024-10-07T08:32:30Z-
dc.date.available2024-10-07T08:32:30Z-
dc.identifier.urihttp://hdl.handle.net/10397/109379-
dc.language.isoenen_US
dc.publisherSpringeren_US
dc.subjectAnatomical structureen_US
dc.subjectContrastive learningen_US
dc.subjectMedical vision-languageen_US
dc.subjectPre-trainingen_US
dc.subjectRepresentation learningen_US
dc.titleAnatomical structure-guided medical vision-language pre-trainingen_US
dc.typeConference Paperen_US
dc.identifier.spage80en_US
dc.identifier.epage90en_US
dc.identifier.doi10.1007/978-3-031-72120-5_8en_US
dcterms.abstractLearning medical visual representations through vision-language pre-training has reached remarkable progress. Despite the promising performance, it still faces challenges, i.e., local alignment lacks interpretability and clinical relevance, and the insufficient internal and external representation learning of image-report pairs. To address these issues, we propose an Anatomical Structure-Guided (ASG) framework. Specifically, we parse raw reports into triplets <anatomical region, finding, existence>, and fully utilize each element as supervision to enhance representation learning. For anatomical region, we design an automatic anatomical region-sentence alignment paradigm in collaboration with radiologists, considering them as the minimum semantic units to explore fine-grained local alignment. For finding and existence, we regard them as image tags, applying an image-tag recognition decoder to associate image features with their respective tags within each sample and constructing soft labels for contrastive learning to improve the semantic association of different image-report pairs. We evaluate the proposed ASG framework on two downstream tasks, including five public benchmarks. Experimental results demonstrate that our method outperforms the state-of-the-art methods.en_US
dcterms.accessRightsembargoed accessen_US
dcterms.bibliographicCitationIn MG Linguraru,Q Dou, A Feragen, S Giannarou, B Glocker, K Lekadir, & JA Schnabel [Eds.]. Medical Image Computing and Computer Assisted Intervention– MICCAI 2024 27th International Conference Marrakesh, Morocco, October 6–10, 2024 Proceedings, Part XI, p. 80-90. Cham, Switzerland: Springer, 2024en_US
dcterms.issued2024-
dc.relation.ispartofbookMedical Image Computing and Computer Assisted Intervention– MICCAI 2024 : 27th International Conference Marrakesh, Morocco, October 6–10, 2024 Proceedings, Part XIen_US
dc.relation.conferenceMedical Image Computing and Computer Assisted Intervention [MICCAI]en_US
dc.description.validate202410 bcchen_US
dc.description.oaNot applicableen_US
dc.identifier.FolderNumbera3073a-
dc.identifier.SubFormID49382-
dc.description.fundingSourceOthersen_US
dc.description.fundingTextStart-up Fund of The Hong Kong Polytechnic University (No. P0045999); the Seed Fund of the Research Institute for Smart Ageing (No. P0050946)en_US
dc.description.pubStatusPublisheden_US
dc.date.embargo2025-10-03en_US
dc.description.oaCategoryGreen (AAM)en_US
Appears in Collections:Conference Paper
Open Access Information
Status embargoed access
Embargo End Date 2025-10-03
Access
View full-text via PolyU eLinks SFX Query
Show simple item record

Page views

64
Citations as of Apr 14, 2025

Google ScholarTM

Check

Altmetric


Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.