Awesome
Are Language Models Actually Useful for Time Series Forecasting? (NeurIPS 2024 Spotlight)
๐ Papar Link
In this work we showed that despite the recent popularity of LLMs in time series forecasting (TSF) they do not appear to meaningfully improve performance. A simple baseline, "PAttn," was proposed, which outperformed most LLM-based TSF models.
Nevertheless, we do not deny the significant potential of LLMs in time series, especially in reasoning about time series through text. You may refer to this time series reasoning work.
Authors: Mingtian Tan,Mike A. Merrill,Vinayak Gupta,Tim Althoff,Thomas Hartvigsen
Overview ๐๐ผ
Recent work in time series analysis has increasingly focused on adapting pretrained large language models (LLMs) for forecasting (TSF), classification, and anomaly detection. These studies suggest that language models, designed for sequential dependencies in text, could generalize to time series data. While this idea aligns with the popularity of language models in machine learning, direct connections between language modeling and TSF remain unclear. How beneficial are language models for traditional TSF task?
Through a series of ablation studies on three recent LLM-based TSF methods, we found that removing the LLM component or replacing it with a simple attention layer did not worsen resultsโin many cases, it even led to improvements. Additionally, we introduced PAttn, showing that patching and attention structures can perform as well as state-of-the-art LLM-based forecasters.
Dataset ๐
You can access the well pre-processed datasets from Google Drive, then place the downloaded contents under ./datasets
Setup ๐ง
Three different popular LLM-based TSF methods were included in our ablation approach. You might want to follow the corresponding repos, OneFitsAll, Time-LLM, and CALF, to set up the environment respectivly. For the ''PAttn'' method, the environment from any of the above repos is compatible.
PAttn ๐ค
The main difference between PAttn and PatchTST is that we gradually removed parts of the Transformer module that may not be as essential, and Position Embedding. For more explanation, please refer to this response.
Motivation: When DLinear has been surpassed by many new methods, we aim to provide a method based on Patching that is both simple and effective, serving as a simple baseline.
cd ./PAttn
bash ./scripts/ETTh.sh (for ETTh1 & ETTh2)
bash ./scripts/ETTm.sh (for ETTm1 & ETTm2)
bash ./scripts/weather.sh (for Weather)
For other datasets, Please change the script name in above command.
Ablations
Run Ablations on CALF (ETT) :
cd ./CALF
sh scripts/long_term_forecasting/ETTh_GPT2.sh
sh scripts/long_term_forecasting/ETTm_GPT2.sh
sh scripts/long_term_forecasting/traffic.sh
(For other datasets, such as traffic)
Run Ablations on OneFitsAll (ETT) :
cd ./OFA
bash ./script/ETTh_GPT2.sh
bash ./script/ETTm_GPT2.sh
bash ./script/illness.sh
(For other datasets, such as illness)
Run Ablations on Time-LLM (ETT)
cd ./Time-LLM-exp
bash ./scripts/train_script/TimeLLM_ETTh1.sh
bash ./scripts/train_script/TimeLLM_ETTm1.sh
bash ./scripts/train_script/TimeLLM_Weather.sh
(For other datasets, such as Weather)
(To run ablations on other datasets, please change the dataset name as shown in example.)
Acknowledgement
This codebase is built based on the Time-Series-Library. Thanks!
Citation
If you find our work useful, please kindly cite our work as follows:
@inproceedings{tan2024language,
title={Are Language Models Actually Useful for Time Series Forecasting?},
author={Tan, Mingtian and Merrill, Mike A and Gupta, Vinayak and Althoff, Tim and Hartvigsen, Thomas},
booktitle={Neural Information Processing Systems (NeurIPS)},
year={2024}
}