This repository contains the code of ST-LLM "Spatial-Temporal Large Language Model for Traffic Prediction" paper
Traffic prediction, an essential component for intelligent transportation systems, endeavours to use historical data to foresee future traffic features at specific locations. Although existing traffic prediction models often emphasize developing complex neural network structures, their accuracy has not improved. Recently, large language models have shown outstanding capabilities in time series analysis. Differing from existing models, LLMs progress mainly through parameter expansion and extensive pretraining while maintaining their fundamental structures. Motivated by these developments, we propose a Spatial-Temporal Large Language Model (ST-LLM) for traffic prediction. In the ST-LLM, we define timesteps at each location as tokens and design a spatial-temporal embedding to learn the spatial location and global temporal patterns of these tokens. Additionally, we integrate these embeddings by a fusion convolution to each token for a unified spatial-temporal representation. Furthermore, we innovate a partially frozen attention strategy to adapt the LLM to capture global spatial-temporal dependencies for traffic prediction. Comprehensive experiments on real traffic datasets offer evidence that ST-LLM is a powerful spatial-temporal learner that outperforms state-of-the-art models. Notably, the ST-LLM also exhibits robust performance in both few-shot and zero-shot prediction scenarios.
- Python 3.11
- PyTorch 2.1.2
- cuda 11.5
- torchvision 0.8.0
> conda env create -f env_ubuntu.yaml
We provide preprocessed datasets, which you can access here.
If you need the original datasets, please refer to the ESG.
CUDA_VISIBLE_DEVICES=0
nohup python train.py --data taxi_pick --device cuda:0 > your_log_name.log &
If you find our work useful in your research. Please consider giving a star ⭐ and citation 📚.
@inproceedings{liu2024spatial,
title={Spatial-temporal large language model for traffic prediction},
author={Liu, Chenxi and Yang, Sun and Xu, Qianxiong and Li, Zhishuai and Long, Cheng and Li, Ziyue and Zhao, Rui},
booktitle={MDM},
year={2024}
}
Our implementation adapts OFA as the code base and has extensively modified it for our purposes. are grateful to the authors for providing their implementations and related resources.