Please use this identifier to cite or link to this item: http://hdl.handle.net/10397/105495
PIRA download icon_1.1View/Download Full Text
DC FieldValueLanguage
dc.contributorDepartment of Computingen_US
dc.creatorChen, Jen_US
dc.creatorWu, XMen_US
dc.creatorLi, Yen_US
dc.creatorLi, Qen_US
dc.creatorZhan, LMen_US
dc.creatorChung, FLen_US
dc.date.accessioned2024-04-15T07:34:42Z-
dc.date.available2024-04-15T07:34:42Z-
dc.identifier.urihttp://hdl.handle.net/10397/105495-
dc.description34th Conference on Neural Information Processing Systems (NeurIPS 2020), 6-12 December 2020, Onlineen_US
dc.language.isoenen_US
dc.publisherNeurIPSen_US
dc.rightsPosted with permission of the author.en_US
dc.titleA closer look at the training strategy for modern meta-learningen_US
dc.typeConference Paperen_US
dc.identifier.spage396en_US
dc.identifier.epage406en_US
dc.identifier.volume33en_US
dcterms.abstractThe support/query (S/Q) episodic training strategy has been widely used in modern meta-learning algorithms and is believed to improve their generalization ability to test environments. This paper conducts a theoretical investigation of this training strategy on generalization. From a stability perspective, we analyze the generalization error bound of generic meta-learning algorithms trained with such strategy. We show that the S/Q episodic training strategy naturally leads to a counterintuitive generalization bound of O(1/√n), which only depends on the task number n but independent of the inner-task sample size m. Under the common assumption m << n for few-shot learning, the bound of O(1/√n) implies strong generalization guarantees for modern meta-learning algorithms in the few-shot regime. To further explore the influence of training strategies on generalization, we propose a leave-one-out (LOO) training strategy for meta-learning and compare it with S/Q training. Experiments on standard few-shot regression and classification tasks with popular meta-learning algorithms validate our analysis.en_US
dcterms.accessRightsopen accessen_US
dcterms.bibliographicCitationAdvances in neural information processing systems, 2020, v. 33, p. 396-406en_US
dcterms.isPartOfAdvances in neural information processing systemsen_US
dcterms.issued2020-
dc.relation.conferenceConference on Neural Information Processing Systems [NeurIPS]en_US
dc.description.validate202402 bcchen_US
dc.description.oaVersion of Recorden_US
dc.identifier.FolderNumberCOMP-0162-
dc.description.fundingSourceRGCen_US
dc.description.pubStatusPublisheden_US
dc.identifier.OPUS56554682-
dc.description.oaCategoryCopyright retained by authoren_US
Appears in Collections:Conference Paper
Files in This Item:
File Description SizeFormat 
Chen_Closer_Look_Training.pdf387.07 kBAdobe PDFView/Open
Open Access Information
Status open access
File Version Version of Record
Show simple item record

Page views

112
Last Week
6
Last month
Citations as of Nov 9, 2025

Downloads

17
Citations as of Nov 9, 2025

Google ScholarTM

Check


Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.