Please use this identifier to cite or link to this item: http://hdl.handle.net/10397/105512
PIRA download icon_1.1View/Download Full Text
DC FieldValueLanguage
dc.contributorDepartment of Computing-
dc.creatorYang, R-
dc.creatorCao, J-
dc.creatorWen, Z-
dc.creatorWu, Y-
dc.creatorHe, X-
dc.date.accessioned2024-04-15T07:34:47Z-
dc.date.available2024-04-15T07:34:47Z-
dc.identifier.isbn978-1-952148-90-3-
dc.identifier.urihttp://hdl.handle.net/10397/105512-
dc.description2020 Conference on Empirical Methods in Natural Language Processing, 16th-20th November 2020, Onlineen_US
dc.language.isoenen_US
dc.publisherAssociation for Computational Linguistics (ACL)en_US
dc.rights© 2020 Association for Computational Linguisticsen_US
dc.rightsThis publication is licensed on a Creative Commons Attribution 4.0 International License. (https://creativecommons.org/licenses/by/4.0/)en_US
dc.rightsThe following publication Ruosong Yang, Jiannong Cao, Zhiyuan Wen, Youzheng Wu, and Xiaodong He. 2020. Enhancing Automated Essay Scoring Performance via Fine-tuning Pre-trained Language Models with Combination of Regression and Ranking. In Findings of the Association for Computational Linguistics: EMNLP 2020, pages 1560–1569, Online. Association for Computational Linguistics is available at https://doi.org/10.18653/v1/2020.findings-emnlp.141.en_US
dc.titleEnhancing automated essay scoring performance via fine-tuning pre-trained language models with combination of regression and rankingen_US
dc.typeConference Paperen_US
dc.identifier.spage1560-
dc.identifier.epage1569-
dc.identifier.doi10.18653/v1/2020.findings-emnlp.141-
dcterms.abstractAutomated Essay Scoring (AES) is a critical text regression task that automatically assigns scores to essays based on their writing quality. Recently, the performance of sentence prediction tasks has been largely improved by using Pre-trained Language Models via fusing representations from different layers, constructing an auxiliary sentence, using multi-task learning, etc. However, to solve the AES task, previous works utilize shallow neural networks to learn essay representations and constrain calculated scores with regression loss or ranking loss, respectively. Since shallow neural networks trained on limited samples show poor performance to capture deep semantic of texts. And without an accurate scoring function, ranking loss and regression loss measures two different aspects of the calculated scores. To improve AES’s performance, we find a new way to fine-tune pre-trained language models with multiple losses of the same task. In this paper, we propose to utilize a pre-trained language model to learn text representations first. With scores calculated from the representations, mean square error loss and the batch-wise ListNet loss with dynamic weights constrain the scores simultaneously. We utilize Quadratic Weighted Kappa to evaluate our model on the Automated Student Assessment Prize dataset. Our model outperforms not only state-of-the-art neural models near 3 percent but also the latest statistic model. Especially on the two narrative prompts, our model performs much better than all other state-of-the-art models.-
dcterms.accessRightsopen accessen_US
dcterms.bibliographicCitationIn Findings of the Association for Computational Linguistics: EMNLP 2020, p. 1560-1569. Stroudsburg, PA, USA: Association for Computational Linguistics (ACL), 2020-
dcterms.issued2020-
dc.relation.ispartofbookFindings of the Association for Computational Linguistics: EMNLP 2020-
dc.relation.conferenceConference on Empirical Methods in Natural Language Processing [EMNLP]-
dc.description.validate202402 bcch-
dc.description.oaVersion of Recorden_US
dc.identifier.FolderNumberCOMP-0197en_US
dc.description.fundingSourceRGCen_US
dc.description.fundingSourceOthersen_US
dc.description.fundingTextHong Kong Red Swastika Society Tai Po Secondary Schoolen_US
dc.description.pubStatusPublisheden_US
dc.identifier.OPUS54284708en_US
dc.description.oaCategoryCCen_US
Appears in Collections:Conference Paper
Files in This Item:
File Description SizeFormat 
2020.findings-emnlp.141.pdf281.15 kBAdobe PDFView/Open
Open Access Information
Status open access
File Version Version of Record
Access
View full-text via PolyU eLinks SFX Query
Show simple item record

Page views

55
Citations as of Feb 23, 2025

Downloads

13
Citations as of Feb 23, 2025

Google ScholarTM

Check

Altmetric


Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.