UniTime
UniTime: A Language-Empowered Unified Model for Cross-Domain Time Series Forecasting (WWW 2024)
Install / Use
/learn @liuxu77/UniTimeREADME
UniTime
This is the official repository of our WWW 2024 paper UniTime: A Language-Empowered Unified Model for Cross-Domain Time Series Forecasting. UniTime aims to explore the potential of utilizing a unified model for generalization across time series application domains. This code base offers an implementation to facilitate cross-domain time series learning research.
<img src='img/intro.jpg' width='750px'>Getting Started
Requirements
Our experimental environments include Python 3.9, Pytorch 1.13.1 with CUDA 11.6, and transformers 4.31.0. To install all dependencies, please use the below command.
pip install -r requirements.txt
Datasets
The pre-processed datasets can be obtained from the link here. Then you may choose to download all_datasets.zip, place this zip file into the dataset folder, and finally unzip the file.
Running
In general, we use a csv file to indicate the executing tasks (including training and evaluations) during the cross-domain learning process. There are five columns in the file.
(1) Data: the name of a dataset, corresponding to a config file in the folder data_configs.
(2) Prediction: the prediction length.
(3) Train: the indicator for training.
(4) Valid: the indicator for validation.
(5) Test: the indicator for testing.
For example, the below command is used to train one model for the tasks listed in the file execute_list/train_all.csv. Note that the argument max_token_num should be set to a value larger than the combined number of tokens in both language instructions and time series patches.
python run.py --gpu 0 --training_list execute_list/train_all.csv --max_token_num 17
In the case of evaluating the pretrained model, please setting the argument is_training to 0 and specifying the inference tasks via inference_list.
python run.py --gpu 0 --training_list execute_list/train_all.csv --max_token_num 17 --is_training 0 --inference_list execute_list/inference_all.csv
Citation
If you find our work useful in your research, please cite:
@inproceedings{liu2024unitime,
title={UniTime: A Language-Empowered Unified Model for Cross-Domain Time Series Forecasting},
author={Liu, Xu and Hu, Junfeng and Li, Yuan and Diao, Shizhe and Liang, Yuxuan and Hooi, Bryan and Zimmermann, Roger},
booktitle={Proceedings of the ACM Web Conference 2024},
year={2024}
}
Acknowledgement
We appreciate the following github repository for sharing the valuable code base and datasets:
https://github.com/thuml/Time-Series-Library
Related Skills
node-connect
345.9kDiagnose OpenClaw node connection and pairing failures for Android, iOS, and macOS companion apps
frontend-design
106.4kCreate distinctive, production-grade frontend interfaces with high design quality. Use this skill when the user asks to build web components, pages, or applications. Generates creative, polished code that avoids generic AI aesthetics.
openai-whisper-api
345.9kTranscribe audio via OpenAI Audio Transcriptions API (Whisper).
qqbot-media
345.9kQQBot 富媒体收发能力。使用 <qqmedia> 标签,系统根据文件扩展名自动识别类型(图片/语音/视频/文件)。
