Abstract
In recent years, Deep Reinforcement Learning (DRL) has demonstrated significant potential in Energy Management Strategies (EMS) for Fuel Cell Vehicles (FCVs). Among these, the Soft Actor-Critic (SAC) algorithm has garnered widespread attention for its superior performance, yet it exhibits shortcomings in convergence speed, stability, and reward value. Moreover, the effectiveness of SAC in mitigating the degradation of fuel cells and lithium batteries remains limited. Therefore, this paper proposes an improved SAC (I-SAC) algorithm, which incorporates Prioritized Experience Replay (PER) and Self-Adaptive Temperature Control (SATC) to enhance performance and effectively extend the lifespan of fuel cells and lithium batteries. Simulation results show that, compared with EMSs based on Double Deep Q-Network (DDQN) and Deep Deterministic Policy Gradient (DDPG), I-SAC significantly reduces hydrogen consumption under various operating conditions, while reducing the degradation of fuel cells and lithium batteries by up to 10.615% and 34.347%, respectively. This study presents a new efficient and robust EMS solution for FCVs.
Get full access to this article
View all access options for this article.
