KSLD-TNet: Key Sample Location and Distillation Transformer Network for Multi-Step Ahead Prediction in Industrial Processes

The multi-step ahead prediction of crucial quality indicators is the cornerstone for optimizing and controlling industrial processes. The accurate multi-step ahead prediction over long prediction horizons holds great potential for improving production performance in industrial processes. However, ex...

Ausführliche Beschreibung

Gespeichert in:
Bibliographische Detailangaben
Veröffentlicht in:IEEE sensors journal 2023-12, p.1-1
Hauptverfasser: Liu, Diju, Wang, Yalin, Liu, Chenliang, Yuan, Xiaofeng, Wang, Kai
Format: Artikel
Sprache:eng
Schlagworte:
Online-Zugang:Volltext bestellen
Tags: Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
Beschreibung
Zusammenfassung:The multi-step ahead prediction of crucial quality indicators is the cornerstone for optimizing and controlling industrial processes. The accurate multi-step ahead prediction over long prediction horizons holds great potential for improving production performance in industrial processes. However, extracting historical features presents a significant obstacle in achieving this objective. Recent advancements have demonstrated that transformer networks offer a promising technical solution to this challenge. Nevertheless, the lack of a sample simplification mechanism makes deep feature extraction difficult. It requires a lot of computational costs, which makes the traditional transformer network less applicable in industrial processes. To explore strategies to overcome these obstacles and enhance the suitability of transformer networks for effective multi-step ahead prediction, this paper proposes a novel key sample location and distillation transformer network (KSLD-TNet). Specifically, it first locates key samples with strong interactions using the attention score matrix. Then, non-key samples are filtered out layer by layer in the KSLD-TNet encoder-decoder structure. In this way, the number of input samples for each layer can be lowered exponentially, reducing the difficulty and calculation amount of deep feature extraction significantly. It is worth noting that this paper also designs an information storage structure to avoid information loss during the sample distillation process. Two industrial process datasets are utilized to construct extensive experiments to demonstrate the effectiveness of the proposed method.
ISSN:1530-437X
1558-1748
DOI:10.1109/JSEN.2023.3336789