Please use this identifier to cite or link to this item: http://hdl.handle.net/10397/106689
PIRA download icon_1.1View/Download Full Text
DC FieldValueLanguage
dc.contributorDepartment of Chinese and Bilingual Studiesen_US
dc.creatorScaboro, Sen_US
dc.creatorPortelli, Ben_US
dc.creatorChersoni, Een_US
dc.creatorSantus, Een_US
dc.creatorSerra ,Gen_US
dc.date.accessioned2024-06-03T02:11:31Z-
dc.date.available2024-06-03T02:11:31Z-
dc.identifier.issn0950-7051en_US
dc.identifier.urihttp://hdl.handle.net/10397/106689-
dc.language.isoenen_US
dc.publisherElsevier BVen_US
dc.rights© 2023 The Authors. Published by Elsevier B.V. This is an open access article under the CC BY license (http://creativecommons.org/licenses/by/4.0/).en_US
dc.rightsThe following publication Scaboro, S., Portelli, B., Chersoni, E., Santus, E., & Serra, G. (2023). Extensive evaluation of transformer-based architectures for adverse drug events extraction. Knowledge-Based Systems, 275, 110675 is available at https://doi.org/10.1016/j.knosys.2023.110675.en_US
dc.subjectAdverse drug eventsen_US
dc.subjectExtractionen_US
dc.subjectSide effectsen_US
dc.subjectTransformersen_US
dc.titleExtensive evaluation of transformer-based architectures for adverse drug events extractionen_US
dc.typeJournal/Magazine Articleen_US
dc.identifier.volume275en_US
dc.identifier.doi10.1016/j.knosys.2023.110675en_US
dcterms.abstractAdverse Drug Event (ADE) extraction is one of the core tasks in digital pharmacovigilance, especially when applied to informal texts. This task has been addressed by the Natural Language Processing community using large pre-trained language models, such as BERT. Despite the great number of Transformer-based architectures used in the literature, it is unclear which of them has better performances and why. Therefore, in this paper we perform an extensive evaluation and analysis of 19 Transformer-based models for ADE extraction on informal texts. We compare the performance of all the considered models on two datasets with increasing levels of informality (forums posts and tweets). We also combine the purely Transformer-based models with two commonly-used additional processing layers (CRF and LSTM), and analyze their effect on the models performance. Furthermore, we use a well-established feature importance technique (SHAP) to correlate the performance of the models with a set of features that describe them: model category (AutoEncoding, AutoRegressive, Text-to-Text), pre-training domain, training from scratch, and model size in number of parameters. At the end of our analyses, we identify a list of take-home messages that can be derived from the experimental data.en_US
dcterms.accessRightsopen accessen_US
dcterms.bibliographicCitationKnowledge-based systems, 5 Sept 2023, v. 275, 110675en_US
dcterms.isPartOfKnowledge-based systemsen_US
dcterms.issued2023-09-
dc.identifier.scopus2-s2.0-85162167418-
dc.identifier.eissn1872-7409en_US
dc.identifier.artn110675en_US
dc.description.validate202405 bcchen_US
dc.description.oaVersion of Recorden_US
dc.identifier.FolderNumbera2727a-
dc.identifier.SubFormID48134-
dc.description.fundingSourceSelf-fundeden_US
dc.description.pubStatusPublisheden_US
dc.description.oaCategoryCCen_US
dc.relation.rdatahttps://github.com/AilabUdineGit/ade-detection-surveyen_US
Appears in Collections:Journal/Magazine Article
Files in This Item:
File Description SizeFormat 
1-s2.0-S0950705123004252-main.pdf1.89 MBAdobe PDFView/Open
Open Access Information
Status open access
File Version Version of Record
Access
View full-text via PolyU eLinks SFX Query
Show simple item record

Page views

2
Citations as of Jun 30, 2024

Downloads

3
Citations as of Jun 30, 2024

SCOPUSTM   
Citations

4
Citations as of Jun 21, 2024

WEB OF SCIENCETM
Citations

3
Citations as of Jun 27, 2024

Google ScholarTM

Check

Altmetric


Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.