Risk-aware Energy Management of Extended Range Electric Delivery Vehicles with Implicit Quantile Network [conference paper]

Conference

IEEE 16th International Conference on Automation Science and Engineering (CASE) - August 20, 2020

Authors

Pengyue Wang, Yan Li (Ph.D. student), Shashi Shekhar (professor), William F Northrop

Abstract

Model-free reinforcement learning (RL) algorithms are used to solve sequential decision-making problems under uncertainty. They are data-driven methods and do not require an explicit model of the studied system or environment. Because of this characteristic, they are widely utilized in Intelligent Transportation Systems (ITS), as real-world transportation systems are highly complex and extremely difficult to model. However, in most literature, decisions are made according to the expected long-term return estimated by the RL algorithm, ignoring the underlying risk. In this work, a distributional RL algorithm called implicit quantile network is adapted for the energy management problem of a delivery vehicle. Instead of only estimating the expected long-term return, the full return distribution is estimated implicitly. This is highly beneficial for applications in ITS, as uncertainty and randomness are intrinsic characteristics of transportation systems. In addition, risk-aware strategies are integrated into the algorithm with the risk measure of conditional value at risk. In this study, we demonstrate that by changing a hyperparameter, the trade-off between fuel efficiency and the risk of running out of battery power during a delivery trip can be controlled according to different application scenarios and personal preferences.

Link to full paper

Risk-aware Energy Management of Extended Range Electric Delivery Vehicles with Implicit Quantile Network

Keywords

data mining

Share