Practicability analysis of online deep reinforcement learning towards energy management strategy of 4WD-BEVs driven by dual-motor in-wheel motors

Deep reinforcement learning (DRL) has emerged as a promising approach for optimizing energy management strategies (EMS) in new energy vehicles. Nevertheless, existing studies typically focus on evaluating the performance of a particular algorithm, overlooking critical implementation details and lack...

Ausführliche Beschreibung

Gespeichert in:
Bibliographische Detailangaben
Veröffentlicht in:Energy (Oxford) 2024-03, Vol.290, p.130123, Article 130123
Hauptverfasser: Feng, Zhiyan, Zhang, Qingang, Zhang, Yiming, Fei, Liangyu, Jiang, Fei, Zhao, Shengdun
Format: Artikel
Sprache:eng
Schlagworte:
Online-Zugang:Volltext
Tags: Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
Beschreibung
Zusammenfassung:Deep reinforcement learning (DRL) has emerged as a promising approach for optimizing energy management strategies (EMS) in new energy vehicles. Nevertheless, existing studies typically focus on evaluating the performance of a particular algorithm, overlooking critical implementation details and lacking comprehensive analysis of their real-world applicability. In this paper, we mitigate this issue by conducting a thorough practicability analysis of DRL-based EMS methods. First, we theoretically analyze the benefits and limitations of existing DRL-based EMS approaches based on taxonomy towards their practicability. Subsequently, a novel EMS method that leverages model-based DRL algorithms that other researchers typically underestimate is proposed. The method newly introduces an uncertainty-aware model-based algorithm known as Probabilistic Ensembles with Trajectory Sampling (PETS) and is validated utilizing a four-wheel-drive (4WD) battery electric vehicle (BEV). After that, we conduct a comprehensive practicability analysis of three state-of-the-art DRL algorithms considering critical aspects for real-world deployment, e.g., hyperparameter sensitivity and algorithm transferability. The results demonstrate that even though the on-policy DRL achieves better asymptotic rewards and the off-policy DRL possesses better convergence, the proposed model-based DRL, PETS-based EMS, outperforms others regarding superior robustness and promising transferability across different extents of relevance between tasks. Besides, energy consumption results demonstrate that the model-based EMS can achieve a considerable 96.6% optimality compared to the baseline dynamic programming (DP). Thus, motivated by the challenges of applying DRL algorithms to real-world EMS, our systematic investigation and new insights contribute to advancing the practical employment of DRL-based EMS for BEVs. •A practicability analysis of DRL-based EMS is conducted.•A novel EMS method leveraging a model-based DRL algorithm, PETS, is proposed.•The sensitivity of the DRL-based EMS to hyperparameters is studied.•The transferability between tasks of the DRL-based EMS is investigated.•EMS based on model-based DRL performs better robustness.
ISSN:0360-5442
DOI:10.1016/j.energy.2023.130123