Please use this identifier to cite or link to this item: http://hdl.handle.net/10397/89095
PIRA download icon_1.1View/Download Full Text
DC FieldValueLanguage
dc.contributorDepartment of Computing-
dc.creatorChen, Y-
dc.creatorMa, Y-
dc.creatorMao, X-
dc.creatorLi, Q-
dc.date.accessioned2021-02-04T02:39:18Z-
dc.date.available2021-02-04T02:39:18Z-
dc.identifier.issn2364-1185-
dc.identifier.urihttp://hdl.handle.net/10397/89095-
dc.language.isoenen_US
dc.publisherSpringerOpenen_US
dc.rights© The Author(s) 2019en_US
dc.rightsOpen Access This article is distributed under the terms of the Creative Commons Attribution 4.0 International License (http://creativecommons.org/licenses/by/4.0/), which permits unrestricted use, distribution, and reproduction in any medium, provided you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons license, and indicate if changes were made.en_US
dc.rightsThe following publication Chen, Y., Ma, Y., Mao, X., & Li, Q. (2019). Multi-task learning for abstractive and extractive summarization. Data Science and Engineering, 4(1), 14-23 is available at https://dx.doi.org/10.1007/s41019-019-0087-7en_US
dc.subjectAttention mechanismen_US
dc.subjectAutomatic document summarizationen_US
dc.subjectMulti-Task learningen_US
dc.titleMulti-task learning for abstractive and extractive summarizationen_US
dc.typeJournal/Magazine Articleen_US
dc.identifier.spage14-
dc.identifier.epage23-
dc.identifier.volume4-
dc.identifier.issue1-
dc.identifier.doi10.1007/s41019-019-0087-7-
dcterms.abstractThe abstractive method and extractive method are two main approaches for automatic document summarization. In this paper, to fully integrate the relatedness and advantages of both approaches, we propose a general unified framework for abstractive summarization which incorporates extractive summarization as an auxiliary task. In particular, our framework is composed of a shared hierarchical document encoder, a hierarchical attention mechanism-based decoder, and an extractor. We adopt multi-task learning method to train these two tasks jointly, which enables the shared encoder to better capture the semantics of the document. Moreover, as our main task is abstractive summarization, we constrain the attention learned in the abstractive task with the labels of the extractive task to strengthen the consistency between the two tasks. Experiments on the CNN/DailyMail dataset demonstrate that both the auxiliary task and the attention constraint contribute to improve the performance significantly, and our model is comparable to the state-of-the-art abstractive models. In addition, we cut half number of labels of the extractive task, pretrain the extractor, and jointly train the two tasks using the estimated sentence salience of the extractive task to constrain the attention of the abstractive task. The results do not decrease much compared with using full-labeled data of the auxiliary task.-
dcterms.accessRightsopen accessen_US
dcterms.bibliographicCitationData science and engineering, Mar. 2019, v. 4, no. 1, p. 14-23-
dcterms.isPartOfData science and engineering-
dcterms.issued2019-03-
dc.identifier.scopus2-s2.0-85065029539-
dc.identifier.eissn2364-1541-
dc.description.validate202101 bcrc-
dc.description.oaVersion of Recorden_US
dc.identifier.FolderNumberOA_Scopus/WOSen_US
dc.description.pubStatusPublisheden_US
dc.description.oaCategoryCCen_US
Appears in Collections:Journal/Magazine Article
Files in This Item:
File Description SizeFormat 
Chen2019_Article_Multi-TaskLearningForAbstracti.pdf1.63 MBAdobe PDFView/Open
Open Access Information
Status open access
File Version Version of Record
Access
View full-text via PolyU eLinks SFX Query
Show simple item record

Page views

108
Last Week
0
Last month
Citations as of Nov 9, 2025

Downloads

37
Citations as of Nov 9, 2025

SCOPUSTM   
Citations

45
Citations as of Dec 19, 2025

Google ScholarTM

Check

Altmetric


Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.