Deep Reinforcement Learning-Based Energy Management for Heavy Duty HEV Considering Discrete-Continuous Hybrid Action Space

  • Zemin Eitan Liu
  • , Yanfei Li*
  • , Quan Zhou
  • , Yong Li
  • , Bin Shuai
  • , Hongming Xu
  • , Min Hua
  • , Guikun Tan
  • , Lubing Xu
  • *Corresponding author for this work

Research output: Contribution to journalArticlepeer-review

Abstract

To reduce the fuel consumption (FC) of heavy duty logistic vehicles (HDLVs), P2 parallel hybridization is a promising solution, and deep reinforcement learning (DRL) is a promising method to optimize energy management strategies (EMSs). However, the complicated discrete-continuous hybrid action space lying in the Psystem presents a challenge to achieve real-time optimal control. Thus, this article proposes a novel DRL algorithm combining auto-tune soft actor-critic (ATSAC) with ordinal regression to optimize the engine torque output and gear shifting simultaneously. ATSAC can adjust the update frequency and learning rate of SAC automatically to improve the generalization, and ordinal regression can convert discrete variables into samplings in continuous space to handle the hybrid action. Moreover, a multidimensional scenario-oriented driving cycle (SODC) is established through naturalistic driving big data (NDBD) as the training cycle to further improve the EMS generalization. By comprehensive comparison with the widely used twin-delayed deep deterministic policy gradient (TD3)-based EMSs, ATSAC achieves significant improvement with 53.70% higher computational efficiency and 12.31% lower negative total reward (NTR) in the training process. Application analysis in unseen real-world driving scenarios shows that only ATSAC-based EMS can obtain real-time optimal control in the testing process. Furthermore, the EMS trained through SODC obtains 81.73% lower NTR than the standard China world transient vehicle cycle (CWTVC), which demonstrates that SODC can represent the real-world driving scenarios much more accurately than CWTVC, especially in low-speed high-load conditions, which are crucial for HDLVs.

Original languageEnglish
Pages (from-to)9864-9876
Number of pages13
JournalIEEE Transactions on Transportation Electrification
Volume10
Issue number4
Early online date7 Feb 2024
DOIs
Publication statusPublished - 27 Dec 2024

Bibliographical note

Publisher Copyright:
© 2015 IEEE.

Keywords

  • Deep reinforcement learning (DRL)
  • energy management strategy (EMS)
  • hybrid action space
  • hybrid electric vehicles (HEVs)
  • Markov chain
  • synthetic driving cycle

ASJC Scopus subject areas

  • Automotive Engineering
  • Transportation
  • Energy Engineering and Power Technology
  • Electrical and Electronic Engineering

Fingerprint

Dive into the research topics of 'Deep Reinforcement Learning-Based Energy Management for Heavy Duty HEV Considering Discrete-Continuous Hybrid Action Space'. Together they form a unique fingerprint.

Cite this