Long-term Transformer and Adaptive Fourier Transform for Dynamic Graph Convolutional Traffic Flow Prediction Study
-
摘要: 针对交通流长期趋势性与非平稳性不易有效建模,以及交通流的隐藏动态时空特征难以捕捉的问题,该文提出一种基于长期Transformer和自适应傅里叶变换的动态图卷积交通流预测模型(ADGformer)。其中,长期门控卷积层通过掩码子序列Transformer从长历史序列中学习压缩的、上下文信息丰富的子序列表示,并利用膨胀门控卷积从子序列的时间表示中有效捕捉交通流的长期趋势特征。并设计一种动态图构造器生成动态可学习图,并利用可学习动态图卷积对节点间潜在的和时变的空间依赖关系进行建模以有效捕获交通流的动态隐藏空间特征。其次,自适应频谱块利用傅里叶变换来增强特征表示并捕获长短期的交互作用,同时通过自适应阈值处理来降低交通流的非平稳性。实验结果表明,所提ADGformer模型具有较好的预测性能。
-
关键词:
- 交通流预测 /
- 动态图卷积 /
- 长期Transformer /
- 傅里叶变换 /
- 非平稳性
Abstract:Objective The rapid development of Intelligent Transportation Systems (ITS), combined with population growth and increasing vehicle ownership, has intensified traffic congestion and road capacity limitations. Accurate and timely traffic flow prediction supports traffic control centers in road management and assists travelers in avoiding congestion and improving travel efficiency, thereby relieving traffic pressure and enhancing road network utilization. However, current models struggle to capture the complex, non-stationary, and long-term dependent characteristics of traffic systems, limiting prediction accuracy. To address these limitations, deep learning approaches—such as dynamic graph convolution and Transformer architectures—have gained prominence in traffic prediction. This study proposes a dynamic graph convolutional traffic flow prediction model (ADGformer) that integrates a long-term Transformer with an adaptive Fourier transform. The model is designed to more effectively capture long-term temporal dependencies, handle non-stationary patterns, and extract latent dynamic spatiotemporal features from traffic flow data. Methods The ADGformer model, based on a Long-Term Transformer and Adaptive Fourier Transform, comprises three primary components:a stacked Long-term Gated Convolutional (LGConv) layer, a spatial convolutional layer, and a Multifaceted Fusion Module. The LGConv layer integrates a Masked Sub-series Transformer and dilated gated convolution to extract long-term trend features from temporally segmented traffic flow series, thereby enhancing long-term prediction performance. The Masked Sub-series Transformer—comprising the Subsequent Temporal Representation Learner (STRL) and a self-supervised task head—learns compact and context-rich representations from extended time series sub-segments. The spatial convolutional layer incorporates a dynamic graph constructor that generates learnable graphs based on the hidden states of spatially related nodes. These dynamic graphs are then processed using learnable dynamic graph convolution to extract latent spatial features that evolve over time. To address the non-stationarity of traffic flow sequences, an adaptive spectral block based on Fourier transform is introduced. Results and Discussions The ADGformer model adaptively models inter-node relationships through a dynamic graph constructor, effectively capturing spatial dependencies within traffic networks and the evolving spatial characteristics of traffic flows. It also learns compressed, context-rich subsequence representations and long-term temporal patterns from extended input sequences. The adaptive spectral block, based on the Fourier transform, reduces the non-stationarity of traffic flow data. To evaluate model performance, this study conducts comparison and ablation experiments on three benchmark datasets. On the PEMSD4 dataset, the ADGformer model reduces MAE under Horizon 3 by approximately 21.57% and 16.72% compared with traditional models such as VAR and LR, respectively. Under Horizon 12 on the England dataset, ADGformer reduces RMSE by 3.35% and 2.44% compared with GWNet and MTGNN, respectively. ADGformer achieves the highest prediction accuracy across all three datasets ( Table 2 ). Visualization comparisons with XGBoost, ASTGCN, and DCRNN on PEMSD4 and England datasets further confirm its superior long-term predictive performance (Fig. 5 ). The model maintains robust performance as the prediction horizon increases. To assess the contributions of individual components, ablation experiments are performed for Horizons 3, 6, and 12. On Horizon 12 of the PEMSD4 dataset, ADGformer improves MAE by approximately 3.69%, 5.09%, 0.92%, and 7.59% relative to the variant models NMS-Trans, NLGConv, NASB, and NDGConv, respectively. On the England dataset, MAPE is reduced by 3.56%, 18.53%, 2.98%, and 4.87%, respectively (Table 3 ). Visual results of the ablation study (Fig. 6 ) show that ADGformer consistently outperforms its variants as the forecast step increases, confirming the effectiveness of each module in the proposed model.Conclusions To address the limitations of existing traffic flow prediction models—namely, (1) failure to capture hidden dynamic spatio-temporal correlations among road network nodes, (2) insufficient learning of key temporal information and long-term trend features from extended historical sequences, and (3) limited ability to reduce the non-stationarity of traffic data—this study proposes a combinatorial traffic flow prediction model, ADGformer, which enables accurate long-term traffic flow prediction. ADGformer employs a Masked Sub-series Transformer to pre-train long historical sequences and extract compact, information-rich subsequence representations. A gated mechanism combined with one-dimensional dilated convolution is then applied to capture long-term trend features and enhance temporal modeling capacity. The model also integrates a spatial convolutional layer constructed from dynamic traffic patterns. Through learnable dynamic graph convolution, it effectively captures evolving spatial dependencies across the network. Furthermore, an adaptive spectral block based on Fourier transform is designed to enhance feature representation and reduce non-stationarity by applying adaptive thresholding to suppress noise, while preserving both long- and short-term interactions. Given the complex and variable nature of traffic flow, future work will consider integrating additional features such as periodicity to further improve prediction performance. -
表 1 数据集详细信息
数据集 节点 边 时间步长(min) 时间跨度(月) PEMSD4 307 16 969 5 2 PEMSD8 170 17 833 5 2 England 314 17 353 15 6 表 2 各种模型在3个数据集上的预测结果
数据集 模型 Horizon 3 Horizon 6 Horizon 12 MAE RMSE MAPE MAE RMSE MAPE MAE RMSE MAPE PEMSD4 HA 3.567 9 6.778 7 0.080 3 3.568 9 6.779 7 0.080 4 3.570 7 6.781 5 0.080 4 VAR 1.662 4 3.0882 0.032 7 2.125 0 4.015 6 0.043 3 2.568 7 4.834 7 0.053 6 XGBoost 1.447 1 2.973 2 0.028 8 1.915 9 4.09
下载: