Revisiting LLMs as Zero-Shot Time-Series Forecasters: Small Noise Can Break Large Models

Cited 0 time in webofscience Cited 0 time in scopus
  • Hit : 41
  • Download : 0
Large Language Models (LLMs) have shown remarkable performance across diverse tasks without domain-specific training, fueling interest in their potential for time-series forecasting. While LLMs have shown potential in zero-shot forecasting through prompting alone, recent studies suggest that LLMs lack inherent effectiveness in forecasting. Given these conflicting findings, a rigorous validation is essential for drawing reliable conclusions. In this paper, we evaluate the effectiveness of LLMs as zero-shot forecasters compared to state-of-the-art domain-specific models. Our experiments show that LLM-based zero-shot forecasters often struggle to achieve high accuracy due to their sensitivity to noise, underperforming even simple domain-specific models. We have explored solutions to reduce LLMs’ sensitivity to noise in the zero-shot setting, but improving their robustness remains a significant challenge. Our findings suggest that rather than emphasizing zero-shot forecasting, a more promising direction would be to focus on fine-tuning LLMs to better process numerical sequences. Our experimental code is available at https://github.com/junwoopark92/ revisiting-LLMs-zeroshot-forecaster.
Publisher
Association for Computational Linguistics (ACL)
Issue Date
2025-07-27
Language
English
Citation

63rd Annual Meeting of the Association for Computational Linguistics, ACL 2025, pp.906 - 922

DOI
10.18653/v1/2025.acl-short.71
URI
http://hdl.handle.net/10203/337139
Appears in Collection
AI-Conference Papers(학술대회논문)
Files in This Item
There are no files associated with this item.

qr_code

  • mendeley

    citeulike


rss_1.0 rss_2.0 atom_1.0