Abstract
An online updating framework of an energy management system (EMS) for a multimode hybrid electric powertrain is proposed via cooperation between the asynchronous advantage actor–critic (A3C)-based deep reinforcement learning (DRL) agent and the Markov chain model (MCM). In the overall framework, the DRL agent periodically updates the energy management policy. The MCM expedites the policy update process by generating plenty of probable future drive cycles using recent historical driving data and supplying them to the training process. Assisted with the MCM, the proposed A3C-based energy management framework can yield near-optimal policy for any type of unknown drive cycle in the recent future. Two types of unknown drive cycles are chosen to demonstrate the efficacy of the proposed framework. Type I unknown drive cycle is also generated from the same recent historical driving data but was not included in the training dataset. Type II drive cycle is neither known to the framework nor generated from the same historical data. In type I unknown drive cycle, the trained A3C-based EMS achieves 99% of the fuel economy obtained by the global-optimal EMS and 0.12% deviation from charge sustainability. The trained A3C-based EMS consumes 6%–12% more fuel than the global-optimal EMS for type II unknown drive cycles.
Talk to us
Join us for a 30 min session where you can share your feedback and ask us any queries you have
More From: IEEE Transactions on Transportation Electrification
Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.