Home

Awesome

ETSformer: Exponential Smoothing Transformers for Time-series Forecasting

<p align="center"> <img src=".\pics\etsformer.png" width = "700" alt="" align=center /> <br><br> <b>Figure 1.</b> Overall ETSformer Architecture. </p>

Official PyTorch code repository for the ETSformer paper. Check out our blog post!

Requirements

  1. Install Python 3.8, and the required dependencies.
  2. Required dependencies can be installed by: pip install -r requirements.txt

Data

Usage

  1. Install the required dependencies.
  2. Download data as above, and place them in the folder, dataset/.
  3. Train the model. We provide the experiment scripts of all benchmarks under the folder ./scripts, e.g. ./scripts/ETTm2.sh. You might have to change permissions on the script files by runningchmod u+x scripts/*.
  4. The script for grid search is also provided, and can be run by ./grid_search.sh.

Acknowledgements

The implementation of ETSformer relies on resources from the following codebases and repositories, we thank the original authors for open-sourcing their work.

Citation

Please consider citing if you find this code useful to your research.

<pre>@article{woo2022etsformer, title={ETSformer: Exponential Smoothing Transformers for Time-series Forecasting}, author={Gerald Woo and Chenghao Liu and Doyen Sahoo and Akshat Kumar and Steven C. H. Hoi}, year={2022}, url={https://arxiv.org/abs/2202.01381}, }</pre>