Please use this identifier to cite or link to this item: http://hdl.handle.net/10397/109116
PIRA download icon_1.1View/Download Full Text
Title: O²-Bert : two-stage target-based sentiment analysis
Authors: Yan, Y
Zhang, BW
Ding, G
Li, W 
Zhang, J
Li, JJ
Gao, W
Issue Date: Jan-2024
Source: Cognitive computation, Jan. 2024, v. 16, no. 1, p. 158-176
Abstract: Target-based sentiment analysis (TBSA) is one of the most important NLP research topics for widespread applications. However, the task is challenging, especially when the targets contain multiple words or do not exist in the sequences. Conventional approaches cannot accurately extract the (target, sentiment) pairs due to the limitations of the fixed end-to-end architecture design. In this paper, we propose a framework named O2-Bert, which consists of Opinion target extraction (OTE-Bert) and Opinion sentiment classification (OSC-Bert) to complete the task in two stages. More specifically, we divide the OTE-Bert into three modules. First, an entity number prediction module predicts the number of entities in a sequence, even in an extreme situation where no entities are contained. Afterwards, with predicted number of entities, an entity starting annotation module is responsible for predicting their starting positions. Finally, an entity length prediction module predicts the lengths of these entities, and thus, accomplishes target extraction. In OSC-Bert, the sentiment polarities of extracted targets from OTE-Bert. According to the characteristics of BERT encoders, our framework can be adapted to short English sequences without domain limitations. For other languages, our approach might work through altering the tokenization. Experimental results on the SemEval 2014-16 benchmarks show that the proposed model achieves competitive performances on both domains (restaurants and laptops) and both tasks (target extraction and sentiment classification), with F1-score as evaluated metrics. Specifically, OTE-Bert achieves 84.63%, 89.20%, 83.16%, and 86.88% F1 scores for target extraction, while OSCBert achieves 82.90%, 80.73%, 76.94%, and 83.58% F1 scores for sentiment classification, on the chosen benchmarks. The statistics validate the effectiveness and robustness of our approach and the new “two-stage paradigm”. In future work, we will explore more possibilities of the new paradigm on other NLP tasks.
Keywords: Entity length prediction
Entity number prediction
Entity starting annotation
O2-Bert
OSC-Bert
OTE-Bert
Publisher: Springer New York LLC
Journal: Cognitive computation 
ISSN: 1866-9956
EISSN: 1866-9964
DOI: 10.1007/s12559-023-10191-y
Rights: © The Author(s) 2023, corrected publication 2023
This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article's Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article's Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/.
The following publication Yan, Y., Zhang, BW., Ding, G. et al. O2-Bert: Two-Stage Target-Based Sentiment Analysis. Cogn Comput 16, 158–176 (2024) is available at https://doi.org/10.1007/s12559-023-10191-y.
Appears in Collections:Journal/Magazine Article

Files in This Item:
File Description SizeFormat 
s12559-023-10191-y.pdf5.33 MBAdobe PDFView/Open
Open Access Information
Status open access
File Version Version of Record
Access
View full-text via PolyU eLinks SFX Query
Show full item record

Page views

22
Citations as of Nov 24, 2024

Downloads

6
Citations as of Nov 24, 2024

Google ScholarTM

Check

Altmetric


Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.