Please use this identifier to cite or link to this item: http://hdl.handle.net/10397/114015
PIRA download icon_1.1View/Download Full Text
DC FieldValueLanguage
dc.contributorDepartment of Data Science and Artificial Intelligenceen_US
dc.creatorTian, Hen_US
dc.creatorLiu, Fen_US
dc.creatorZhou, Zen_US
dc.creatorLiu, Ten_US
dc.creatorZhang, Cen_US
dc.creatorHan, Ben_US
dc.date.accessioned2025-07-10T01:31:38Z-
dc.date.available2025-07-10T01:31:38Z-
dc.identifier.urihttp://hdl.handle.net/10397/114015-
dc.descriptionNeurIPS 2024: The Thirty-Eighth Annual Conference on Neural Information Processing Systems, Vancouver, 10-15 Dec 2024en_US
dc.language.isoenen_US
dc.publisherNeurIPSen_US
dc.rightsPosted with permission of the author.en_US
dc.titleMind the gap between prototypes and images in cross-domain finetuningen_US
dc.typeConference Paperen_US
dc.identifier.volume37en_US
dcterms.abstractIn cross-domain few-shot classification (CFC), recent works mainly focus on adapting a simple transformation head on top of a frozen pre-trained backbone with few labeled data to project embeddings into a task-specific metric space where classification can be performed by measuring similarities between image instance and prototype representations. Technically, an assumption implicitly adopted in such a framework is that the prototype and image instance embeddings share the same representation transformation. However, in this paper, we find that there naturally exists a gap, which resembles the modality gap, between the prototype and image instance embeddings extracted from the frozen pre-trained backbone, and simply applying the same transformation during the adaptation phase constrains exploring the optimal representation distributions and shrinks the gap between prototype and image representations. To solve this problem, we propose a simple yet effective method, contrastive prototype-image adaptation (CoPA), to adapt different transformations for prototypes and images similarly to CLIP by treating prototypes as text prompts. Extensive experiments on Meta-Dataset demonstrate that CoPA achieves the state-of-the-art performance more efficiently. Meanwhile, further analyses also indicate that CoPA can learn better representation clusters, enlarge the gap, and achieve the minimum validation loss at the enlarged gap.en_US
dcterms.accessRightsopen accessen_US
dcterms.bibliographicCitationAdvances in neural information processing systems, 2024, v. 37en_US
dcterms.isPartOfAdvances in neural information processing systemsen_US
dcterms.issued2024-
dc.relation.conferenceConference on Neural Information Processing Systems [NeurIPS]en_US
dc.description.validate202507 bcwhen_US
dc.description.oaVersion of Recorden_US
dc.identifier.FolderNumbera3866-
dc.identifier.SubFormID51472-
dc.description.fundingSourceSelf-fundeden_US
dc.description.pubStatusPublisheden_US
dc.description.oaCategoryCopyright retained by authoren_US
Appears in Collections:Conference Paper
Files in This Item:
File Description SizeFormat 
Tian_Mind_Gap_Prototypes.pdf2.02 MBAdobe PDFView/Open
Open Access Information
Status open access
File Version Version of Record
Show simple item record

Google ScholarTM

Check


Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.