Home

Awesome

Learning Deep Time-index Models for Time Series Forecasting (ICML 2023)

<p align="center"> <img src=".\pics\deeptime.png" width = "700" alt="" align=center /> <br><br> <b>Figure 1.</b> Overall approach of DeepTime. </p>

Official PyTorch code repository for the DeepTime paper. Check out our blog post!

Requirements

Dependencies for this project can be installed by:

pip install -r requirements.txt

Quick Start

Data

To get started, you will need to download the datasets as described in our paper:

Reproducing Experiment Results

We provide some scripts to quickly reproduce the results reported in our paper. There are two options, to run the full hyperparameter search, or to directly run the experiments with hyperparameters provided in the configuration files.

Option A: Run the full hyperparameter search.

  1. Run the following command to generate the experiments: make build-all path=experiments/configs/hp_search.
  2. Run the following script to perform training and evaluation: ./run_hp_search.sh (you may need to run chmod u+x run_hp_search.sh first).

Option B: Directly run the experiments with hyperparameters provided in the configuration files.

  1. Run the following command to generate the experiments: make build-all path=experiments/configs/ETTm2.
  2. Run the following script to perform training and evaluation: ./run.sh (you may need to run chmod u+x run.sh first).

Finally, results can be viewed on tensorboard by running tensorboard --logdir storage/experiments/, or in the storage/experiments/experiment_name/metrics.npy file.

Main Results

We conduct extensive experiments on both synthetic and real world datasets, showing that DeepTime has extremely competitive performance, achieving state-of-the-art results on 20 out of 24 settings for the multivariate forecasting benchmark based on MSE.

<p align="center"> <img src=".\pics\results.png" width = "700" alt="" align=center /> <br><br> </p>

Detailed Usage

Further details of the code repository can be found here. The codebase is structured to generate experiments from a .gin configuration file based on the build.variables_dict argument.

  1. First, build the experiment from a config file. We provide 2 ways to build an experiment.
    1. Build a single config file:
      make build config=experiments/configs/folder_name/file_name.gin
      
    2. Build a group of config files:
      make build-all path=experiments/configs/folder_name
      
  2. Next, run the experiment using the following command
    python -m experiments.forecast --config_path=storage/experiments/experiment_name/config.gin run
    
    Alternatively, the first step generates a command file found in storage/experiments/experiment_name/command, which you can use by the following command,
    make run command=storage/experiments/experiment_name/command
    
  3. Finally, you can observe the results on tensorboard
    tensorboard --logdir storage/experiments/
    
    or view the storage/experiments/deeptime/experiment_name/metrics.npy file.

Acknowledgements

The implementation of DeepTime relies on resources from the following codebases and repositories, we thank the original authors for open-sourcing their work.

Citation

Please consider citing if you find this code useful to your research.

<pre>@InProceedings{pmlr-v202-woo23b, title = {Learning Deep Time-index Models for Time Series Forecasting}, author = {Woo, Gerald and Liu, Chenghao and Sahoo, Doyen and Kumar, Akshat and Hoi, Steven}, booktitle = {Proceedings of the 40th International Conference on Machine Learning}, pages = {37217--37237}, year = {2023}, editor = {Krause, Andreas and Brunskill, Emma and Cho, Kyunghyun and Engelhardt, Barbara and Sabato, Sivan and Scarlett, Jonathan}, volume = {202}, series = {Proceedings of Machine Learning Research}, month = {23--29 Jul}, publisher = {PMLR}, pdf = {https://proceedings.mlr.press/v202/woo23b/woo23b.pdf}, url = {https://proceedings.mlr.press/v202/woo23b.html} }</pre>