ASPDD: An Adaptive Knowledge Distillation Framework for TSP Generalization Problems

The traveling salesman problem (TSP) is a classic non-deterministic polynomial-hard (NP-hard) problem. Currently, almost all the research works utilizing Transformers to solve TSP problems employ supervised learning. However, it is extremely challenging to obtain accurate solution labels for the mod...

Ausführliche Beschreibung

Gespeichert in:
Bibliographische Detailangaben
Veröffentlicht in:IEEE access 2024, Vol.12, p.52902-52910
Hauptverfasser: Zheng, Sisi, Ye, Rongye
Format: Artikel
Sprache:eng
Schlagworte:
Online-Zugang:Volltext
Tags: Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
Beschreibung
Zusammenfassung:The traveling salesman problem (TSP) is a classic non-deterministic polynomial-hard (NP-hard) problem. Currently, almost all the research works utilizing Transformers to solve TSP problems employ supervised learning. However, it is extremely challenging to obtain accurate solution labels for the model in large-scale instances, resulting in a severe lack of scale generalization capability. Recent research combines knowledge distillation and Transformer to effectively address the distribution generalization issue. Nonetheless, if the framework is directly applied to the problem of scale generalization, the solution is not satisfactory. To address the aforementioned issues, we propose an adaptive soft probability distributed distillation (ASPDD) framework to improve Transformer scale generalization capability. The ASPDD framework uses a soft probability distributed distillation (SPDD) method to improve the knowledge interaction between the student and teacher models. In particular, ASPDD introduces an adaptive selection strategy, so that the student model can find weak points for improvement training in each training. This framework is utilized for training a model for a small-scale instance (TSP20) and deploying it to a large-scale instance. Extensive benchmarking (10000 instances) demonstrates that our ASPDD framework can achieve competitive results as compared to other Transformer baseline models and knowledge distillation frameworks. In addition, the ASPDD framework is applied to eleven publicly accessible benchmark datasets (TSPLIB). On six benchmark datasets, the experimental results demonstrate that our ASPDD framework outperforms previous knowledge distillation models.
ISSN:2169-3536
2169-3536
DOI:10.1109/ACCESS.2024.3387851