Abstract
With the advancement of autonomous driving and vehicular networking technologies, platoon lane-changing (PLC) has become a research hotspot in intelligent transportation systems. This paper proposes a cooperative control model based on an end-to-end multi-agent deep meta-reinforcement learning (MADMRL) framework to address the technical challenges in PLC. The model considers the coupling effects among vehicles within the platoon, enabling precise control of longitudinal acceleration and lateral front-wheel steering angles. To improve training efficiency and learning outcomes, meta-learning is integrated with platoon dynamics models, proposing the Platoon-MMAPPO algorithm, which enhances model accuracy and accelerates policy network convergence. Additionally, a Platoon-Adaptive-Weight Reward Function (Platoon-Ada-Weight RF) is proposed to effectively guide the learning process, reduce unnecessary exploration, and accelerate the convergence to optimal policies. Highway simulation experiments and ablation studies validate the proposed model’s significant advantages in improving lane-changing efficiency, driving comfort, reducing road occupancy, and ensuring safety.
Get full access to this article
View all access options for this article.
