TY - JOUR
T1 - Cost-Efficient Power Orchestration in Electric Vehicle-Integrated VPP Using Lightweight Multi-Agent Reinforcement Learning
AU - Zheng, Dongyu
AU - Liu, Lei
AU - Tang, Guoming
AU - Guo, Deke
AU - Hu, Jia
AU - Khowaja, Sunder Ali
AU - Dev, Kapal
N1 - Publisher Copyright:
© 2000-2011 IEEE.
PY - 2025
Y1 - 2025
N2 - Virtual power plants (VPPs) have emerged as an advanced solution for coordinating distributed energy resources (DERs), including the stored energy of electric vehicles (EVs). The substantial demand for EV charging imposes significant stress on the electrical grid, resulting in elevated energy costs for operators. On the other hand, the advent of reversible charging technologies offers a promising method to harness the surplus energy from EVs that do not require immediate charging. In this study, we introduce the concept of EV-integrated VPP in place of the traditional charging station. By designing a tailored mathematical model, we optimize the charging and discharging schedule, termed optimal power orchestration, which aims to minimize the energy costs as well as EV battery degradation. We further design a lightweight multi-agent reinforcement learning (MARL) based approach to tackle the optimal power orchestration problem by reformulating it as a decentralized partially observable Markov decision process (Dec-POMDP). Meanwhile, knowledge distillation is also incorporated into the proposed method to enable efficient deployment in such a distributed resource-constrained environment. Through extensive experiments utilizing real-world EV charging data and realistic scenario settings, our findings demonstrate significant reductions in energy costs and battery degradation by 15.5% and 71.1%, respectively, compared to the baseline method.
AB - Virtual power plants (VPPs) have emerged as an advanced solution for coordinating distributed energy resources (DERs), including the stored energy of electric vehicles (EVs). The substantial demand for EV charging imposes significant stress on the electrical grid, resulting in elevated energy costs for operators. On the other hand, the advent of reversible charging technologies offers a promising method to harness the surplus energy from EVs that do not require immediate charging. In this study, we introduce the concept of EV-integrated VPP in place of the traditional charging station. By designing a tailored mathematical model, we optimize the charging and discharging schedule, termed optimal power orchestration, which aims to minimize the energy costs as well as EV battery degradation. We further design a lightweight multi-agent reinforcement learning (MARL) based approach to tackle the optimal power orchestration problem by reformulating it as a decentralized partially observable Markov decision process (Dec-POMDP). Meanwhile, knowledge distillation is also incorporated into the proposed method to enable efficient deployment in such a distributed resource-constrained environment. Through extensive experiments utilizing real-world EV charging data and realistic scenario settings, our findings demonstrate significant reductions in energy costs and battery degradation by 15.5% and 71.1%, respectively, compared to the baseline method.
KW - demand charge
KW - knowledge distillation
KW - reinforcement learning
KW - VPP
UR - https://www.scopus.com/pages/publications/105009437016
U2 - 10.1109/TITS.2025.3577438
DO - 10.1109/TITS.2025.3577438
M3 - Article
AN - SCOPUS:105009437016
SN - 1524-9050
JO - IEEE Transactions on Intelligent Transportation Systems
JF - IEEE Transactions on Intelligent Transportation Systems
ER -