TY - JOUR
T1 - Fuel-Efficient Switching Control for Platooning Systems With Deep Reinforcement Learning
AU - Goncalves, Tiago Rocha
AU - Cunha, Rafael Fernandes
AU - Varma, Vineeth Satheeskumar
AU - Elayoubi, Salah Eddine
N1 - Publisher Copyright:
© 2000-2011 IEEE.
PY - 2023/12/1
Y1 - 2023/12/1
N2 - The wide appeal of fuel-efficient transport solutions is constantly increasing due to the major impact of the transportation industry on the environment. Platooning systems represent a relatively simple approach in terms of deployment toward fuel-efficient solutions. This paper addresses the reduction of fuel consumption in platooning systems attainable by dynamically switching between two control policies: Adaptive Cruise Control (ACC) and Cooperative Adaptive Cruise Control (CACC). The switching rule is dictated by a Deep Reinforcement Learning (DRL) technique to overcome unpredictable platoon disturbances and to learn appropriate transient shift times while maximizing fuel efficiency. However, due to safety and convergence issues of DRL, our algorithm establishes transition times and minimum periods of operation of ACC and CACC controllers instead of directly controlling vehicles. Numerical experiments show that the DRL agent outperforms both static ACC and CACC versions and the threshold logic control in terms of fuel efficiency while also being robust to perturbations and satisfying safety requirements.
AB - The wide appeal of fuel-efficient transport solutions is constantly increasing due to the major impact of the transportation industry on the environment. Platooning systems represent a relatively simple approach in terms of deployment toward fuel-efficient solutions. This paper addresses the reduction of fuel consumption in platooning systems attainable by dynamically switching between two control policies: Adaptive Cruise Control (ACC) and Cooperative Adaptive Cruise Control (CACC). The switching rule is dictated by a Deep Reinforcement Learning (DRL) technique to overcome unpredictable platoon disturbances and to learn appropriate transient shift times while maximizing fuel efficiency. However, due to safety and convergence issues of DRL, our algorithm establishes transition times and minimum periods of operation of ACC and CACC controllers instead of directly controlling vehicles. Numerical experiments show that the DRL agent outperforms both static ACC and CACC versions and the threshold logic control in terms of fuel efficiency while also being robust to perturbations and satisfying safety requirements.
KW - cooperative adaptive cruise control (CACC)
KW - deep reinforcement learning
KW - Vehicle platoons
UR - http://www.scopus.com/inward/record.url?scp=85168685735&partnerID=8YFLogxK
U2 - 10.1109/TITS.2023.3304977
DO - 10.1109/TITS.2023.3304977
M3 - Article
AN - SCOPUS:85168685735
SN - 1524-9050
VL - 24
SP - 13989
EP - 13999
JO - IEEE Transactions on Intelligent Transportation Systems
JF - IEEE Transactions on Intelligent Transportation Systems
IS - 12
ER -