Please use this identifier to cite or link to this item: http://hdl.handle.net/10397/97881
PIRA download icon_1.1View/Download Full Text
DC FieldValueLanguage
dc.contributorDepartment of Chinese and Bilingual Studiesen_US
dc.creatorScaboro, Sen_US
dc.creatorPortelli, Ben_US
dc.creatorChersoni, Een_US
dc.creatorSantus, Een_US
dc.creatorSerra, Gen_US
dc.date.accessioned2023-03-24T07:39:47Z-
dc.date.available2023-03-24T07:39:47Z-
dc.identifier.isbn978-1-954085-90-9en_US
dc.identifier.urihttp://hdl.handle.net/10397/97881-
dc.descriptionThe Seventh Workshop on Noisy User-generated Text (W-NUT 2021), Nov 11, 2021, Onlineen_US
dc.language.isoenen_US
dc.publisherAssociation for Computational Linguisticsen_US
dc.rights©2021 Association for Computational Linguisticsen_US
dc.rightsMaterials published in or after 2016 are licensed on a Creative Commons Attribution 4.0 International License. (https://creativecommons.org/licenses/by/4.0/)en_US
dc.rightsThe following publication Simone Scaboro, Beatrice Portelli, Emmanuele Chersoni, Enrico Santus, and Giuseppe Serra. 2021. NADE: A Benchmark for Robust Adverse Drug Events Extraction in Face of Negations. In Proceedings of the Seventh Workshop on Noisy User-generated Text (W-NUT 2021), pages 230–237, Online. Association for Computational Linguistics. is available at https://aclanthology.org/2021.wnut-1.26.pdf.en_US
dc.titleNADE : a benchmark for robust adverse drug events extraction in face of negationsen_US
dc.typeConference Paperen_US
dc.identifier.spage230en_US
dc.identifier.epage237en_US
dc.identifier.doi10.18653/v1/2021.wnut-1.26en_US
dcterms.abstractAdverse Drug Event (ADE) extraction models can rapidly examine large collections of social media texts, detecting mentions of drug-related adverse reactions and trigger medical investigations. However, despite the recent advances in NLP, it is currently unknown if such models are robust in face of negation, which is pervasive across language varieties. In this paper we evaluate three state-of-the-art systems, showing their fragility against negation, and then we introduce two possible strategies to increase the robustness of these models: a pipeline approach, relying on a specific component for negation detection; an augmentation of an ADE extraction dataset to artificially create negated samples and further train the models. We show that both strategies bring significant increases in performance, lowering the number of spurious entities predicted by the models. Our dataset and code will be publicly released to encourage research on the topic.en_US
dcterms.accessRightsopen accessen_US
dcterms.bibliographicCitationIn W. Xu, A. Ritter, T. Baldwin & A. Rahimi (Eds.), Proceedings of the Seventh Workshop on Noisy User-generated Text (W-NUT 2021), p. 230-237, Online. Association for Computational Linguistics, 2021en_US
dcterms.issued2021-11-
dc.relation.conferenceWorkshop on Noisy User-generated Text [W-NUT]en_US
dc.description.validate202303 bcwwen_US
dc.description.oaVersion of Recorden_US
dc.identifier.FolderNumberCBS-0071-
dc.description.fundingSourceSelf-fundeden_US
dc.description.pubStatusPublisheden_US
dc.identifier.OPUS56021339-
Appears in Collections:Conference Paper
Files in This Item:
File Description SizeFormat 
2021.wnut-1.26.pdf298.53 kBAdobe PDFView/Open
Open Access Information
Status open access
File Version Version of Record
Access
View full-text via PolyU eLinks SFX Query
Show simple item record

Page views

64
Citations as of May 5, 2024

Downloads

20
Citations as of May 5, 2024

Google ScholarTM

Check

Altmetric


Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.