Volt-VAR Control in Active Distribution Networks Using Multi-Agent Reinforcement Learning
Abstract
:1. Introduction
- A method based on electrical distance is proposed for partitioning DNs. The partitioned DN exhibits highly aggregated characteristics within regions and low coupling between regions, laying the foundation for achieving distributed VVC of PV inverters.
- This paper proposes a framework for VVC in DNs based on the Multi-Agent Soft Actor-Critic (MASAC) algorithm. The framework employs a strategy of centralized training followed by distributed execution, aiming to reduce communication and computation demands on the DNs during execution. This approach alleviates the resource-intensive nature of centralized control strategies in terms of real-time computation and storage requirements. The established framework enables the coordinated control of PV systems to minimize voltage deviations while simultaneously minimizing reactive power losses in the DN. Importantly, this coordination occurs with agents interacting only with local information from sub-regions of the DN.
- In order to validate the effectiveness and versatility of the proposed framework, experiments were conducted using five MARL algorithms, including MASAC, on IEEE 33-bus and IEEE 141-bus network. The results demonstrate that the proposed method can effectively achieve VVC in DNs, relying solely on local observation information after training.
2. Proposed VVC Strategy
2.1. The Model of Volt-VAR Control
2.2. Network Partition Based on Electrical Distance
2.3. Formulate VVC as a Markov Game
2.4. MASAC-Based VVC Framework
3. Case Study
3.1. The Performance of Network Partition
3.2. The Performance of Volt-VAR Control Based on MASAC
- The counterfactual multi-agent (COMA) method, which uses Q-value decomposition to achieve collaborative decision-making by optimizing local Q-values and influence factors;
- The MADDPG algorithm, which uses a deep deterministic policy gradient to achieve collaborative decision-making for multi-agents;
- The multi-agent proximal policy optimization (MAPPO) algorithm, which uses policy optimization and importance sampling to achieve collaborative decision-making for multi-agents;
- Multi-agent twin delayed deep deterministic (MATD3) method, which uses double-delayed deep deterministic policy gradients to achieve collaborative decision-making for multi-agents;
- The proposed MASAC method, where SAC algorithm-based agents are trained based on global observations, and each agent controls the inverter reactive output based on local observations within the network in which it is located.
- Rewards: this metric calculates the value of discount rewards received by the agent after executing an action. The agent aims to maximize the discounted rewards, and a higher discounted reward indicates that in the current episode, the agent receives a higher value of the reward from the environment after executing the scheduling action and has a better overall performance in maximizing the trade-off between reducing the voltage excursion and reducing the reactive power loss.
- Controllable ratio: this metric calculates the ratio of time steps during which all bus voltages are under control during each episode. A higher controllable ratio indicates a better performance of the algorithm in terms of bus control.
- Reactive power loss: this metric calculates the average value of the reactive power loss of all lines for each time step during each event. A lower reactive power loss can indicate that the algorithm has better performance in reducing power loss.
3.2.1. Test on This IEEE 33-Bus Network
3.2.2. Test on the IEEE 141-Bus Network
4. Conclusions
- Broadening the scope of controllable power electronic devices;
- Improving the scalability of the algorithm to tackle voltage and VVC challenges in larger-scale DNs.
Author Contributions
Funding
Data Availability Statement
Conflicts of Interest
References
- Wang, Y.; Qiu, D.; Strbac, G.; Gao, Z. Coordinated Electric Vehicle Active and Reactive Power Control for Active Distribution Networks. IEEE Trans. Ind. Inform. 2022, 19, 1611–1622. [Google Scholar] [CrossRef]
- Liu, H.; Wu, W. Online Multi-Agent Reinforcement Learning for Decentralized Inverter-Based Volt-VAR Control. IEEE Trans. Smart Grid 2021, 12, 2980–2990. [Google Scholar] [CrossRef]
- Lee, X.Y.; Sarkar, S.; Wang, Y. A Graph Policy Network Approach for Volt-Var Control in Power Distribution Systems. Appl. Energy 2022, 323, 119530. [Google Scholar] [CrossRef]
- Dorostkar-Ghamsari, M.R.; Fotuhi-Firuzabad, M.; Lehtonen, M.; Safdarian, A. Value of Distribution Network Reconfiguration in Presence of Renewable Energy Resources. IEEE Trans. Power Syst. 2016, 31, 1879–1888. [Google Scholar] [CrossRef]
- Zhang, L.; Shen, C.; Chen, Y.; Huang, S.; Tang, W. Coordinated Allocation of Distributed Generation, Capacitor Banks and Soft Open Points in Active Distribution Networks Considering Dispatching Results. Appl. Energy 2018, 231, 1122–1131. [Google Scholar] [CrossRef]
- Long, C.; Ochoa, L.F. Voltage Control of PV-Rich LV Networks: OLTC-Fitted Transformer and Capacitor Banks. IEEE Trans. Power Syst. 2016, 31, 4016–4025. [Google Scholar] [CrossRef]
- Minetti, M.; Rosini, A.; Denegri, G.B.; Bonfiglio, A.; Procopio, R. An Advanced Droop Control Strategy for Reactive Power Assessment in Islanded Microgrids. IEEE Trans. Power Syst. 2021, 37, 3014–3025. [Google Scholar] [CrossRef]
- Antoniadou-Plytaria, K.E.; Kouveliotis-Lysikatos, I.N.; Georgilakis, P.S.; Hatziargyriou, N.D. Distributed and Decentralized Voltage Control of Smart Distribution Networks: Models, Methods, and Future Research. IEEE Trans. Smart Grid 2017, 8, 2999–3008. [Google Scholar] [CrossRef]
- Li, Z.; Su, S.; Jin, X.; Xia, M.; Chen, Q.; Yamashita, K. Stochastic and Distributed Optimal Energy Management of Active Distribution Networks within Integrated Office Buildings. CSEE J. Power Energy Syst. 2024, 10, 504–517. [Google Scholar] [CrossRef]
- Xu, T.; Wu, W. Accelerated ADMM-Based Fully Distributed Inverter-Based Volt/Var Control Strategy for Active Distribution Networks. IEEE Trans. Ind. Inform. 2020, 16, 7532–7543. [Google Scholar] [CrossRef]
- Li, P.; Zhang, C.; Wu, Z.; Xu, Y.; Hu, M.; Dong, Z. Distributed Adaptive Robust Voltage/Var Control with Network Partition in Active Distribution Networks. IEEE Trans. Smart Grid 2019, 11, 2245–2256. [Google Scholar] [CrossRef]
- Ma, W.; Wang, W.; Chen, Z.; Hu, R. A Centralized Voltage Regulation Method for Distribution Networks Containing High Penetrations of Photovoltaic Power. Int. J. Electr. Power Energy Syst. 2021, 129, 106852. [Google Scholar] [CrossRef]
- Kumar Tatikayala, V.; Dixit, S. Multi-Stage Voltage Control in High Photovoltaic Based Distributed Generation Penetrated Distribution System Considering Smart Inverter Reactive Power Capability. Ain Shams Eng. J. 2024, 15, 102265. [Google Scholar] [CrossRef]
- Yang, T.; Guo, Y.; Deng, L.; Sun, H.; Wu, W. A Linear Branch Flow Model for Radial Distribution Networks and Its Application to Reactive Power Optimization and Network Reconfiguration. IEEE Trans. Smart Grid 2021, 12, 2027–2036. [Google Scholar] [CrossRef]
- Liu, H.; Zhang, C.; Guo, Q. Data-Driven Robust Voltage/VAR Control Using PV Inverters in Active Distribution Networks. In Proceedings of the 2020 International Conference on Smart Grids and Energy Systems (SGES), Perth, Australia, 23–26 November 2020; pp. 314–319. [Google Scholar]
- Sun, X.; Qiu, J. Two-Stage Volt/Var Control in Active Distribution Networks with Multi-Agent Deep Reinforcement Learning Method. IEEE Trans. Smart Grid 2021, 12, 2903–2912. [Google Scholar] [CrossRef]
- Cao, D.; Zhao, J.; Hu, W.; Ding, F.; Huang, Q.; Chen, Z.; Blaabjerg, F. Data-Driven Multi-Agent Deep Reinforcement Learning for Distribution System Decentralized Voltage Control With High Penetration of PVs. IEEE Trans. Smart Grid 2021, 12, 4137–4150. [Google Scholar] [CrossRef]
- Chen, Y.; Liu, Y.; Zhao, J.; Qiu, G.; Yin, H.; Li, Z. Physical-Assisted Multi-Agent Graph Reinforcement Learning Enabled Fast Voltage Regulation for PV-Rich Active Distribution Network. Appl. Energy 2023, 351, 121743. [Google Scholar] [CrossRef]
- Thurner, L.; Scheidler, A.; Schäfer, F.; Menke, J.-H.; Dollichon, J.; Meier, F.; Meinecke, S.; Braun, M. Pandapower—An Open-Source Python Tool for Convenient Modeling, Analysis, and Optimization of Electric Power Systems. IEEE Trans. Power Syst. 2018, 33, 6510–6521. [Google Scholar] [CrossRef]
- Zimmerman, R.D.; Murillo-Sánchez, C.E.; Thomas, R.J. MATPOWER: Steady-State Operations, Planning, and Analysis Tools for Power Systems Research and Education. IEEE Trans. Power Syst. 2011, 26, 12–19. [Google Scholar] [CrossRef]
Networks | Capacity | Location |
---|---|---|
IEEE 33 | 0.5MW/0.51MVA | 11, 17, 21, 24, 29, 32 |
IEEE 141 | 0.5MW/0.51MVA | 35, 52, 58, 61, 67, 68, 74, 76, 81, 86, 99, 105, 109, 110, 115, 116, 129, 132, 136, 137, 138, 140 |
Parameter | Value |
---|---|
h | 0.8 |
Batch | 32 |
Experience replay buffer | 5000 |
Policy network learning rate | 0.001 |
Critic network learning rate | 0.001 |
λ | 0.99 |
Strategy | Average Voltage | Maximum Voltage Rise | Minimum Voltage Drop | Reactive Power Losses | Controllable Ratio |
---|---|---|---|---|---|
MASAC | 0.9976 | 0.002 | 0.001 | 0.1233 | 95.37% |
MADDPG | 0.9944 | 0.004 | 0.002 | 0.1706 | 96.60% |
MATD3 | 0.9879 | 0.002 | 0.001 | 0.0899 | 96.03% |
MAPPO | 1.0096 | 0.010 | 0.005 | 0.6310 | 60.92% |
COMA | 0.9972 | 0.001 | 0.003 | 0.2223 | 98.85% |
Strategy | Average Voltage | Maximum Voltage Rise | Minimum Voltage Drop | Reactive Power Losses | Controllable Ratio |
---|---|---|---|---|---|
MASAC | 0.9957 | 0.003 | 0.001 | 0.5653 | 98.41% |
MADDPG | 0.9932 | 0.006 | 0.002 | 0.7159 | 98.50% |
MATD3 | 0.9854 | 0.005 | 0.001 | 0.7940 | 98.23% |
MAPPO | 1.0255 | 0.013 | 0.004 | 1.0755 | 68.73% |
COMA | 1.0046 | 0.001 | 0.003 | 1.1765 | 99.99% |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2024 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Su, S.; Zhan, H.; Zhang, L.; Xie, Q.; Si, R.; Dai, Y.; Gao, T.; Wu, L.; Zhang, J.; Shang, L. Volt-VAR Control in Active Distribution Networks Using Multi-Agent Reinforcement Learning. Electronics 2024, 13, 1971. https://doi.org/10.3390/electronics13101971
Su S, Zhan H, Zhang L, Xie Q, Si R, Dai Y, Gao T, Wu L, Zhang J, Shang L. Volt-VAR Control in Active Distribution Networks Using Multi-Agent Reinforcement Learning. Electronics. 2024; 13(10):1971. https://doi.org/10.3390/electronics13101971
Chicago/Turabian StyleSu, Shi, Haozhe Zhan, Luxi Zhang, Qingyang Xie, Ruiqi Si, Yuxin Dai, Tianlu Gao, Linhan Wu, Jun Zhang, and Lei Shang. 2024. "Volt-VAR Control in Active Distribution Networks Using Multi-Agent Reinforcement Learning" Electronics 13, no. 10: 1971. https://doi.org/10.3390/electronics13101971
APA StyleSu, S., Zhan, H., Zhang, L., Xie, Q., Si, R., Dai, Y., Gao, T., Wu, L., Zhang, J., & Shang, L. (2024). Volt-VAR Control in Active Distribution Networks Using Multi-Agent Reinforcement Learning. Electronics, 13(10), 1971. https://doi.org/10.3390/electronics13101971