Please use this identifier to cite or link to this item:
http://hdl.handle.net/10397/91924
Title: | Is domain adaptation worth your investment? Comparing BERT and FinBERT on financial tasks | Authors: | Peng, B Chersoni, E Hsu, YY Huang, CR |
Issue Date: | 2021 | Source: | In Proceedings of the Third Workshop on Economics and Natural Language Processing (ECONLP 2021), November 11, 2021, Punta Cana, Dominican Republic and Online, p. 37–44. Stroudsburg, PA: Association for Computational Linguistics (ACL), 2021 | Abstract: | With the recent rise in popularity of Transformer models in Natural Language Processing, research efforts have been dedicated to the development of domain-adapted versions of BERT-like architectures. In this study, we focus on FinBERT, a Transformer model trained on text from the financial domain. By comparing its performances with the original BERT on a wide variety of financial text processing tasks, we found continual pretraining from the original model to be the more beneficial option. Domain-specific pretraining from scratch, conversely, seems to be less effective. |
Publisher: | Association for Computational Linguistics (ACL) | ISBN: | 978-1-954085-84-8 | DOI: | 10.18653/v1/2021.econlp-1.5 | Rights: | ©2021 Association for Computational Linguistics ACL materials are Copyright © 1963–2021 ACL; other materials are copyrighted by their respective copyright holders. Materials prior to 2016 here are licensed under the Creative Commons Attribution-NonCommercial-ShareAlike 3.0 International License. Permission is granted to make copies for the purposes of teaching and research. Materials published in or after 2016 are licensed on a Creative Commons Attribution 4.0 International License (https://creativecommons.org/licenses/by/4.0/). The following publication Peng, B., Chersoni, E., Hsu, Y. Y., & Huang, C. R. (2021, November). Is Domain Adaptation Worth Your Investment? Comparing BERT and FinBERT on Financial Tasks. In Proceedings of the Third Workshop on Economics and Natural Language Processing (pp. 37-44) is available at https://doi.org/10.18653/v1/2021.econlp-1.5 |
Appears in Collections: | Conference Paper |
Files in This Item:
File | Description | Size | Format | |
---|---|---|---|---|
Peng_Domain_Adaptation_Worth.pdf | 396.5 kB | Adobe PDF | View/Open |
Page views
41
Last Week
0
0
Last month
Citations as of Jun 4, 2023
Downloads
15
Citations as of Jun 4, 2023

Google ScholarTM
Check
Altmetric
Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.