Travel order quantity prediction via attention-based bidirectional LSTM networks

Traffic flow prediction is a very challenging task in traffic networks. Travel order quantity prediction is of great value to the analysis of traffic flow. However, the number of travel orders is closely related to time, leading to its short-term proximity and long-term cyclical dependence. Therefor...

Ausführliche Beschreibung

Gespeichert in:
Bibliographische Detailangaben
Veröffentlicht in:The Journal of supercomputing 2022-02, Vol.78 (3), p.4398-4420
Hauptverfasser: Yang, Fei, Zhang, Huyin, Tao, Shiming
Format: Artikel
Sprache:eng
Schlagworte:
Online-Zugang:Volltext
Tags: Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
Beschreibung
Zusammenfassung:Traffic flow prediction is a very challenging task in traffic networks. Travel order quantity prediction is of great value to the analysis of traffic flow. However, the number of travel orders is closely related to time, leading to its short-term proximity and long-term cyclical dependence. Therefore, it is difficult to predict travel order quantity using traditional methods. To capture both the long-term and short-term correlations, in this paper, we propose an attention-based bidirectional long short-term memory network (AT-BLSTM) model, which consists of bidirectional long short-term memory (BLSTM) layer and attention layer. The BLSTM layer contains forward and backward long short-term memory (LSTM), which uses a novel method to combine forward and backward output. The attention layer utilizes novel self-attention algorithm to assign different weights according to the correlation between features. In this case, AT-BLSTM can predict travel order quantity more accurately than other time series models. For example, AT-BLSTM achieves the lowest MAE (0.0647), the lowest RMSE (0.0836) and the lowest MAPE (0.1239) among all the methods on dataset Xiuying. Extensive experiments on real-world travel datasets offer evidence that the proposed approach matches or outperforms state-of-the-art methods.
ISSN:0920-8542
1573-0484
DOI:10.1007/s11227-021-04032-8