Please use this identifier to cite or link to this item:
http://hdl.handle.net/10397/107878
Title: | VIBE : topic-driven temporal adaptation for twitter classification | Authors: | Zhang, Y Li, J Li, W |
Issue Date: | 2023 | Source: | In Proceedings of the 2023 Conference on Empirical Methods in Natural Language Processing, p. 3340–3354, Singapore. Association for Computational Linguistics, 2023 | Abstract: | Language features are evolving in real-world social media, resulting in the deteriorating performance of text classification in dynamics. To address this challenge, we study temporal adaptation, where models trained on past data are tested in the future. Most prior work focused on continued pretraining or knowledge updating, which may compromise their performance on noisy social media data. To tackle this issue, we reflect feature change via modeling latent topic evolution and propose a novel model, VIBE: Variational Information Bottleneck for Evolutions. Concretely, we first employ two Information Bottleneck (IB) regularizers to distinguish past and future topics. Then, the distinguished topics work as adaptive features via multi-task training with timestamp and class label prediction. In adaptive learning, VIBE utilizes retrieved unlabeled data from online streams created posterior to training data time. Substantial Twitter experiments on three classification tasks show that our model, with only 3% of data, significantly outperforms previous state-of-the-art continued-pretraining methods. | Publisher: | Association for Computational Linguistics (ACL) | ISBN: | 979-8-89176-060-8 | Description: | The 2023 Conference on Empirical Methods in Natural Language Processing, December 6-10, 2023, Singapore | Rights: | © 2023 Association for Computational Linguistics Materials published in or after 2016 are licensed on a Creative Commons Attribution 4.0 International License (https://creativecommons.org/licenses/by/4.0/). The following publication Xianming Li and Jing Li. 2024. BeLLM: Backward Dependency Enhanced Large Language Model for Sentence Embeddings. In Proceedings of the 2024 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies (Volume 1: Long Papers), pages 792–804, Mexico City, Mexico. Association for Computational Linguistics is available at https://aclanthology.org/2023.emnlp-main.203/. |
Appears in Collections: | Conference Paper |
Files in This Item:
File | Description | Size | Format | |
---|---|---|---|---|
2023.emnlp-main.203.pdf | 2.11 MB | Adobe PDF | View/Open |
Page views
5
Citations as of Jul 21, 2024
Downloads
1
Citations as of Jul 21, 2024
Google ScholarTM
Check
Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.