Skip to yearly menu bar Skip to main content


Spotlight Poster

Are Language Models Actually Useful for Time Series Forecasting?

Mingtian Tan · Mike Merrill · Vinayak Gupta · Tim Althoff · Tom Hartvigsen

[ ] [ Project Page ]
Fri 13 Dec 4:30 p.m. PST — 7:30 p.m. PST

Abstract:

Large language models (LLMs) are being applied to time series tasks, particularly time series forecasting. However, are language models actually useful for time series? After a series of ablation studies on three recent and popular LLM-based time series forecasting methods, we find that removing the LLM component or replacing it with a basic attention layer does not degrade the forecasting results; in most cases the results even improved. We also find that despite their significant computational cost pretrained LLMs do no better than models trained from scratch, do not represent the sequential dependencies in time series, and do not assist in few-shot settings. Additionally, we explore time series encoders and reveal that patching and attention structures perform similarly to state-of-the-art LLM-based forecasters. All code, environments, and data to reproduce our work are available in this anonymized repository: https://anonymous.4open.science/r/TS_Models-D1BC/

Live content is unavailable. Log in and register to view live content