Please use this identifier to cite or link to this item: http://hdl.handle.net/10397/99805
PIRA download icon_1.1View/Download Full Text
DC FieldValueLanguage
dc.contributorDepartment of Computingen_US
dc.creatorYuan, Ren_US
dc.creatorWang, Zen_US
dc.creatorCao, Zen_US
dc.creatorLi, Wen_US
dc.date.accessioned2023-07-21T01:07:33Z-
dc.date.available2023-07-21T01:07:33Z-
dc.identifier.urihttp://hdl.handle.net/10397/99805-
dc.language.isoenen_US
dc.publisherAssociation for Computational Linguisticsen_US
dc.rights© 2022 Association for Computational Linguisticsen_US
dc.rightsMaterials published in or after 2016 are licensed on a Creative Commons Attribution 4.0 International License. (https://creativecommons.org/licenses/by/4.0/).en_US
dc.rightsThe following publication Ruifeng Yuan, Zili Wang, Ziqiang Cao, and Wenjie Li. 2022. Few-shot Query-Focused Summarization with Prefix-Merging. In Proceedings of the 2022 Conference on Empirical Methods in Natural Language Processing, pages 3704–3714, Abu Dhabi, United Arab Emirates. Association for Computational Linguistics is available at https://aclanthology.org/2022.emnlp-main.243/.en_US
dc.titleFew-shot query-focused summarization with prefix-mergingen_US
dc.typeConference Paperen_US
dc.identifier.spage3704en_US
dc.identifier.epage3714en_US
dcterms.abstractQuery-focused summarization has been considered as an important extension for text summarization. It aims to generate a concise highlight for a given query. Different from text summarization, query-focused summarization has long been plagued by the problem of lacking high-quality large-scale datasets. In this paper, we investigate the idea that whether we can integrate and transfer the knowledge of text summarization and question answering to assist the few-shot learning in query-focused summarization. Here, we propose prefix-merging, a prefix-based pretraining strategy for few-shot learning in query-focused summarization. Drawn inspiration from prefix-tuning, we are allowed to integrate the task knowledge from text summarization and question answering into a properly designed prefix and apply the merged prefix to query-focused summarization. With only a small amount of trainable parameters, prefix-merging outperforms fine-tuning on query-focused summarization. We further discuss the influence of different prefix designs and propose a visualized explanation for how prefix-merging works.en_US
dcterms.accessRightsopen accessen_US
dcterms.bibliographicCitationProceedings of the 2022 Conference on Empirical Methods in Natural Language Processing, December 7-11, 2022, Abu Dhabi, United Arab Emirates, p. 3704 - 3714. Stroudsburg, PA: Association for Computational Linguistics (ACL), 2022en_US
dcterms.issued2022-
dc.relation.conferenceConference on Empirical Methods in Natural Language Processing [EMNLP]en_US
dc.description.validate202307 bcwwen_US
dc.description.oaVersion of Recorden_US
dc.identifier.FolderNumbera2311-
dc.identifier.SubFormID47468-
dc.description.fundingSourceRGCen_US
dc.description.pubStatusPublisheden_US
dc.description.oaCategoryCCen_US
Appears in Collections:Conference Paper
Files in This Item:
File Description SizeFormat 
2022.emnlp-main.243.pdf358.56 kBAdobe PDFView/Open
Open Access Information
Status open access
File Version Version of Record
Show simple item record

Page views

153
Last Week
7
Last month
Citations as of Nov 9, 2025

Downloads

61
Citations as of Nov 9, 2025

Google ScholarTM

Check


Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.