LAPSE:2023.31283
Published Article
LAPSE:2023.31283
Influence of the Reward Function on the Selection of Reinforcement Learning Agents for Hybrid Electric Vehicles Real-Time Control
Matteo Acquarone, Claudio Maino, Daniela Misul, Ezio Spessa, Antonio Mastropietro, Luca Sorrentino, Enrico Busto
April 18, 2023
The real-time control optimization of electrified vehicles is one of the most demanding tasks to be faced in the innovation progress of low-emissions mobility. Intelligent energy management systems represent interesting solutions to solve complex control problems, such as the maximization of the fuel economy of hybrid electric vehicles. In the recent years, reinforcement-learning-based controllers have been shown to outperform well-established real-time strategies for specific applications. Nevertheless, the effects produced by variation in the reward function have not been thoroughly analyzed and the potential of the adoption of a given RL agent under different testing conditions is still to be assessed. In the present paper, the performance of different agents, i.e., Q-learning, deep Q-Network and double deep Q-Network, are investigated considering a full hybrid electric vehicle throughout multiple driving missions and introducing two distinct reward functions. The first function aims at guaranteeing a charge-sustaining policy whilst reducing the fuel consumption (FC) as much as possible; the second function in turn aims at minimizing the fuel consumption whilst ensuring an acceptable battery state of charge (SOC) by the end of the mission. The novelty brought by the results of this paper lies in the demonstration of a non-trivial incapability of DQN and DDQN to outperform traditional Q-learning when a SOC-oriented reward is considered. On the contrary, optimal fuel consumption reductions are attained by DQN and DDQN when more complex FC-oriented minimization is deployed. Such an important outcome is particularly evident when the RL agents are trained on regulatory driving cycles and tested on unknown real-world driving missions.
Keywords
Artificial Intelligence, fuel consumption, hybrid electric vehicles, real-time control, reinforcement learning
Suggested Citation
Acquarone M, Maino C, Misul D, Spessa E, Mastropietro A, Sorrentino L, Busto E. Influence of the Reward Function on the Selection of Reinforcement Learning Agents for Hybrid Electric Vehicles Real-Time Control. (2023). LAPSE:2023.31283
Author Affiliations
Acquarone M: Interdepartmental Center for Automotive Research and Sustainable Mobility (
Maino C: Interdepartmental Center for Automotive Research and Sustainable Mobility ( [ORCID]
Misul D: Interdepartmental Center for Automotive Research and Sustainable Mobility ( [ORCID]
Spessa E: Interdepartmental Center for Automotive Research and Sustainable Mobility ( [ORCID]
Mastropietro A: Department of Data Science, EURECOM, Route des Chappes 450, 06904 Biot, France [ORCID]
Sorrentino L: Addfor Industriale s.r.l., Piazza Solferino 7, 10121 Turin, Italy
Busto E: Addfor Industriale s.r.l., Piazza Solferino 7, 10121 Turin, Italy
Journal Name
Energies
Volume
16
Issue
6
First Page
2749
Year
2023
Publication Date
2023-03-15
Published Version
ISSN
1996-1073
Version Comments
Original Submission
Other Meta
PII: en16062749, Publication Type: Journal Article
Record Map
Published Article

LAPSE:2023.31283
This Record
External Link

doi:10.3390/en16062749
Publisher Version
Download
Files
[Download 1v1.pdf] (5.7 MB)
Apr 18, 2023
Main Article
License
CC BY 4.0
Meta
Record Statistics
Record Views
124
Version History
[v1] (Original Submission)
Apr 18, 2023
 
Verified by curator on
Apr 18, 2023
This Version Number
v1
Citations
Most Recent
This Version
URL Here
https://psecommunity.org/LAPSE:2023.31283
 
Original Submitter
Auto Uploader for LAPSE
Links to Related Works
Directly Related to This Work
Publisher Version