Please use this identifier to cite or link to this item: http://hdl.handle.net/10397/95764
PIRA download icon_1.1View/Download Full Text
DC FieldValueLanguage
dc.contributorDepartment of Computingen_US
dc.creatorLiu, Sen_US
dc.creatorCao, Jen_US
dc.creatorYang, Ren_US
dc.creatorWen, Zen_US
dc.date.accessioned2022-10-06T06:04:23Z-
dc.date.available2022-10-06T06:04:23Z-
dc.identifier.issn0306-4573en_US
dc.identifier.urihttp://hdl.handle.net/10397/95764-
dc.language.isoenen_US
dc.publisherPergamon Pressen_US
dc.rights© 2022 Elsevier Ltd. All rights reserved.en_US
dc.rightsThe following publication Liu, S., Cao, J., Yang, R., & Wen, Z. (2022). Key phrase aware transformer for abstractive summarization. Information Processing & Management, 59(3), 102913 is available at https://dx.doi.org/10.1016/j.ipm.2022.102913.en_US
dc.subjectAbstractive summarizationen_US
dc.subjectDeep learningen_US
dc.subjectKey phrase extractionen_US
dc.subjectText summarizationen_US
dc.titleKey phrase aware transformer for abstractive summarizationen_US
dc.typeJournal/Magazine Articleen_US
dc.identifier.volume59en_US
dc.identifier.issue3en_US
dc.identifier.doi10.1016/j.ipm.2022.102913en_US
dcterms.abstractAbstractive summarization aims to generate a concise summary covering salient content from single or multiple text documents. Many recent abstractive summarization methods are built on the transformer model to capture long-range dependencies in the input text and achieve parallelization. In the transformer encoder, calculating attention weights is a crucial step for encoding input documents. Input documents usually contain some key phrases conveying salient information, and it is important to encode these phrases completely. However, existing transformer-based summarization works did not consider key phrases in input when determining attention weights. Consequently, some of the tokens within key phrases only receive small attention weights, which is not conducive to encoding the semantic information of input documents. In this paper, we introduce some prior knowledge of key phrases into the transformer-based summarization model and guide the model to encode key phrases. For the contextual representation of each token in the key phrase, we assume the tokens within the same key phrase make larger contributions compared with other tokens in the input sequence. Based on this assumption, we propose the Key Phrase Aware Transformer (KPAT), a model with the highlighting mechanism in the encoder to assign greater attention weights for tokens within key phrases. Specifically, we first extract key phrases from the input document and score the phrases’ importance. Then we build the block diagonal highlighting matrix to indicate these phrases’ importance scores and positions. To combine self-attention weights with key phrases’ importance scores, we design two structures of highlighting attention for each head and the multi-head highlighting attention. Experimental results on two datasets (Multi-News and PubMed) from different summarization tasks and domains show that our KPAT model significantly outperforms advanced summarization baselines. We conduct more experiments to analyze the impact of each part of our model on the summarization performance and verify the effectiveness of our proposed highlighting mechanism.en_US
dcterms.accessRightsopen accessen_US
dcterms.bibliographicCitationInformation processing and management, May 2022, v. 59, no. 3, 102913en_US
dcterms.isPartOfInformation processing and managementen_US
dcterms.issued2022-05-
dc.identifier.scopus2-s2.0-85126037956-
dc.identifier.artn102913en_US
dc.description.validate202210 bckwen_US
dc.description.oaAccepted Manuscripten_US
dc.identifier.FolderNumbera1754-
dc.identifier.SubFormID45887-
dc.description.fundingSourceRGCen_US
dc.description.fundingSourceOthersen_US
dc.description.fundingTextHong Kong Jockey Club Charities Trusten_US
dc.description.pubStatusPublisheden_US
dc.description.oaCategoryGreen (AAM)en_US
Appears in Collections:Journal/Magazine Article
Files in This Item:
File Description SizeFormat 
Liu_Multi-Head_Abstractive_Summarization.pdfPre-Published version1.57 MBAdobe PDFView/Open
Open Access Information
Status open access
File Version Final Accepted Manuscript
Access
View full-text via PolyU eLinks SFX Query
Show simple item record

Page views

134
Last Week
1
Last month
Citations as of Nov 10, 2025

Downloads

290
Citations as of Nov 10, 2025

SCOPUSTM   
Citations

36
Citations as of Dec 19, 2025

WEB OF SCIENCETM
Citations

30
Citations as of Dec 18, 2025

Google ScholarTM

Check

Altmetric


Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.