Please use this identifier to cite or link to this item: http://hdl.handle.net/10397/107878
PIRA download icon_1.1View/Download Full Text
Title: VIBE : topic-driven temporal adaptation for twitter classification
Authors: Zhang, Y 
Li, J 
Li, W 
Issue Date: 2023
Source: In Proceedings of the 2023 Conference on Empirical Methods in Natural Language Processing, p. 3340–3354, Singapore. Association for Computational Linguistics, 2023
Abstract: Language features are evolving in real-world social media, resulting in the deteriorating performance of text classification in dynamics. To address this challenge, we study temporal adaptation, where models trained on past data are tested in the future. Most prior work focused on continued pretraining or knowledge updating, which may compromise their performance on noisy social media data. To tackle this issue, we reflect feature change via modeling latent topic evolution and propose a novel model, VIBE: Variational Information Bottleneck for Evolutions. Concretely, we first employ two Information Bottleneck (IB) regularizers to distinguish past and future topics. Then, the distinguished topics work as adaptive features via multi-task training with timestamp and class label prediction. In adaptive learning, VIBE utilizes retrieved unlabeled data from online streams created posterior to training data time. Substantial Twitter experiments on three classification tasks show that our model, with only 3% of data, significantly outperforms previous state-of-the-art continued-pretraining methods.
Publisher: Association for Computational Linguistics (ACL)
ISBN: 979-8-89176-060-8
Description: The 2023 Conference on Empirical Methods in Natural Language Processing, December 6-10, 2023, Singapore
Rights: © 2023 Association for Computational Linguistics
Materials published in or after 2016 are licensed on a Creative Commons Attribution 4.0 International License (https://creativecommons.org/licenses/by/4.0/).
The following publication Xianming Li and Jing Li. 2024. BeLLM: Backward Dependency Enhanced Large Language Model for Sentence Embeddings. In Proceedings of the 2024 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies (Volume 1: Long Papers), pages 792–804, Mexico City, Mexico. Association for Computational Linguistics is available at https://aclanthology.org/2023.emnlp-main.203/.
Appears in Collections:Conference Paper

Files in This Item:
File Description SizeFormat 
2023.emnlp-main.203.pdf2.11 MBAdobe PDFView/Open
Open Access Information
Status open access
File Version Version of Record
Access
View full-text via PolyU eLinks SFX Query
Show full item record

Page views

5
Citations as of Jul 21, 2024

Downloads

1
Citations as of Jul 21, 2024

Google ScholarTM

Check


Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.