Please use this identifier to cite or link to this item:
http://hdl.handle.net/10397/112415
Title: | Helformer : an attention-based deep learning model for cryptocurrency price forecasting | Authors: | Kehinde, TO Adedokun, OJ Joseph, A Kabirat, KM Akano, HA Olanrewaju , OA |
Issue Date: | Dec-2025 | Source: | Journal of big data, Dec. 2025, v. 12, no. 1, 81 | Abstract: | Cryptocurrencies have become a significant asset class, attracting considerable attention from investors and researchers due to their potential for high returns despite inherent price volatility. Traditional forecasting methods often fail to accurately predict price movements as they do not account for the non-linear and non-stationary nature of cryptocurrency data. In response to these challenges, this study introduces the Helformer model, a novel deep learning approach that integrates Holt-Winters exponential smoothing with Transformer-based deep learning architecture. This integration allows for a robust decomposition of time series data into level, trend, and seasonality components, enhancing the model’s ability to capture complex patterns in cryptocurrency markets. To optimize the model’s performance, Bayesian hyperparameter tuning via Optuna, including a pruner callback, was utilized to efficiently find optimal model parameters while reducing training time by early termination of suboptimal training runs. Empirical results from testing the Helformer model against other advanced deep learning models across various cryptocurrencies demonstrate its superior predictive accuracy and robustness. The model not only achieves lower prediction errors but also shows remarkable generalization capabilities across different types of cryptocurrencies. Additionally, the practical applicability of the Helformer model is validated through a trading strategy that significantly outperforms traditional strategies, confirming its potential to provide actionable insights for traders and financial analysts. The findings of this study are particularly beneficial for investors, policymakers, and researchers, offering a reliable tool for navigating the complexities of cryptocurrency markets and making informed decisions. | Keywords: | Bitcoin Cryptocurrency forecasting Helformer Neural networks Time series Transformer |
Publisher: | SpringerOpen | Journal: | Journal of big data | EISSN: | 2196-1115 | DOI: | 10.1186/s40537-025-01135-4 | Rights: | © The Author(s) 2025. Open Access This article is licensed under a Creative Commons Attribution-NonCommercial-NoDerivatives 4.0 International License, which permits any non-commercial use, sharing, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if you modified the licensed material. You do not have permission under this licence to share adapted material derived from this article or parts of it. The images or other third party material in this article are included in the article’s Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article’s Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by-nc-nd/4.0/. The following publication Kehinde, T.O., Adedokun, O.J., Joseph, A. et al. Helformer: an attention-based deep learning model for cryptocurrency price forecasting. J Big Data 12, 81 (2025) is available at https://doi.org/10.1186/s40537-025-01135-4. |
Appears in Collections: | Journal/Magazine Article |
Files in This Item:
File | Description | Size | Format | |
---|---|---|---|---|
s40537-025-01135-4.pdf | 3.75 MB | Adobe PDF | View/Open |
Page views
10
Citations as of Apr 14, 2025
Downloads
4
Citations as of Apr 14, 2025

Google ScholarTM
Check
Altmetric
Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.