Uncertainty-aware Energy Management of Extended Range Electric Delivery Vehicles with Bayesian Ensemble [conference paper]

Conference

IEEE Intelligent Vehicles Symposium - October 19 - November 13, 2020

Authors

Pengyue Wang, Yan Li (Ph.D. student), Shashi Shekhar (professor), William F. Northrop

Abstract

In recent years, deep reinforcement learning (DRL) algorithms have been widely studied and utilized in the area of Intelligent Transportation Systems (ITS). DRL agents are mostly trained with transition pairs and interaction trajectories generated from simulation, and they can achieve satisfying or near optimal performances under familiar input states. However, for relative rare visited or even unvisited regions in the state space, there is no guarantee that the agent could perform well. Unfortunately, novel conditions are inevitable in real-world problems and there is always a gap between the real data and simulated data. Therefore, to implement DRL algorithms in real-world transportation systems, we should not only train the agent learn a policy that maps states to actions, but also the model uncertainty associated with each action. In this study, we adapt the method of Bayesian ensemble to train a group of agents with imposed diversity for an energy management system of a delivery vehicle. The agents in the ensemble agree well on familiar states but show diverse results on unfamiliar or novel states. This uncertainty estimation facilitates the implementation of interpretable postprocessing modules which can ensure robust and safe operations under high uncertainty conditions.

Link to full paper

Uncertainty-aware Energy Management of Extended Range Electric Delivery Vehicles with Bayesian Ensemble

Keywords

intelligent transportation systems

Share