Deep Reinforcement Learning-Based Distribution Network Planning Method Considering Renewable Energy
Liang Ma (),
Chenyi Si,
Ke Wang,
Jinshan Luo,
Shigong Jiang and
Yi Song
Additional contact information
Liang Ma: State Grid Economic and Technological Research Institute Co., Ltd., Beijing 102209, China
Chenyi Si: School of Electrical and Information Engineering, Tianjin University, Tianjin 300072, China
Ke Wang: School of Electrical and Information Engineering, Tianjin University, Tianjin 300072, China
Jinshan Luo: State Grid Economic and Technological Research Institute Co., Ltd., Beijing 102209, China
Shigong Jiang: State Grid Economic and Technological Research Institute Co., Ltd., Beijing 102209, China
Yi Song: State Grid Economic and Technological Research Institute Co., Ltd., Beijing 102209, China
Energies, 2025, vol. 18, issue 5, 1-17
Abstract:
Distribution networks are an indispensable component of modern economic societies. Against the background of building new power systems, the rapid growth of distributed renewable energy sources, such as photovoltaic and wind power, has introduced many challenges for distribution network planning (DNP), including different source-load compositions, complex network topologies, and varied application scenarios. Traditional heuristic algorithms are limited in scalability and struggle to address the increasingly complex optimization problems of DNP. The emergence of new artificial intelligence provides a new way to solve this problem. Based on the above discussion, this paper proposes a DNP method based on deep reinforcement learning (DRL). By defining state space and action space, a Markov decision process model tailored for DNP is formulated. Then, a multi-objective optimization function and a corresponding reward function including construction costs, voltage deviation, renewable energy penetration, and electricity purchase costs are designed to guide the generation of network topology schemes. Based on the proximal policy optimization algorithm, an actor-critic-based autonomous generation and adaptive adjustment model for DNP is constructed. Finally, the representative test case is selected to verify the effectiveness of the proposed method, which indicates that the proposed method can improve the efficiency of DNP and promote the digital transformation of DNP.
Keywords: DNP method; DRL; Markov decision process model; proximal policy optimization (search for similar items in EconPapers)
JEL-codes: Q Q0 Q4 Q40 Q41 Q42 Q43 Q47 Q48 Q49 (search for similar items in EconPapers)
Date: 2025
References: Add references at CitEc
Citations:
Downloads: (external link)
https://www.mdpi.com/1996-1073/18/5/1254/pdf (application/pdf)
https://www.mdpi.com/1996-1073/18/5/1254/ (text/html)
Related works:
This item may be available elsewhere in EconPapers: Search for items with the same title.
Export reference: BibTeX
RIS (EndNote, ProCite, RefMan)
HTML/Text
Persistent link: https://EconPapers.repec.org/RePEc:gam:jeners:v:18:y:2025:i:5:p:1254-:d:1605032
Access Statistics for this article
Energies is currently edited by Ms. Agatha Cao
More articles in Energies from MDPI
Bibliographic data for series maintained by MDPI Indexing Manager ().