EconPapers    
Economics at your fingertips  
 

Research on Dynamic Energy Management Optimization of Park Integrated Energy System Based on Deep Reinforcement Learning

Xinjian Jiang, Lei Zhang, Fuwang Li, Zhiru Li, Zhijian Ling and Zhenghui Zhao ()
Additional contact information
Xinjian Jiang: Shanxi Energy Internet Research Institute, Taiyuan 030000, China
Lei Zhang: Shanxi Energy Internet Research Institute, Taiyuan 030000, China
Fuwang Li: Shanxi Energy Internet Research Institute, Taiyuan 030000, China
Zhiru Li: Shanxi Energy Internet Research Institute, Taiyuan 030000, China
Zhijian Ling: Shanxi Energy Internet Research Institute, Taiyuan 030000, China
Zhenghui Zhao: Shanxi Energy Internet Research Institute, Taiyuan 030000, China

Energies, 2025, vol. 18, issue 19, 1-28

Abstract: Under the background of energy transition, the Integrated Energy System (IES) of the park has become a key carrier for enhancing the consumption capacity of renewable energy due to its multi-energy complementary characteristics. However, the high proportion of wind and solar resource access and the fluctuation of diverse loads have led to the system facing dual uncertainty challenges, and traditional optimization methods are difficult to adapt to the dynamic and complex dispatching requirements. To this end, this paper proposes a new dynamic energy management method based on Deep Reinforcement Learning (DRL) and constructs an IES hybrid integer nonlinear programming model including wind power, photovoltaic, combined heat and power generation, and storage of electric heat energy, with the goal of minimizing the operating cost of the system. By expressing the dispatching process as a Markov decision process, a state space covering wind and solar output, multiple loads and energy storage states is defined, a continuous action space for unit output and energy storage control is constructed, and a reward function integrating economic cost and the penalty for renewable energy consumption is designed. The Deep Deterministic Policy Gradient (DDPG) and Deep Q-Network (DQN) algorithms were adopted to achieve policy optimization. This study is based on simulation rather than experimental validation, which aligns with the exploratory scope of this research. The simulation results show that the DDPG algorithm achieves an average weekly operating cost of 532,424 yuan in the continuous action space scheduling, which is 8.6% lower than that of the DQN algorithm, and the standard deviation of the cost is reduced by 19.5%, indicating better robustness. Under the fluctuation of 10% to 30% on the source-load side, the DQN algorithm still maintains a cost fluctuation of less than 4.5%, highlighting the strong adaptability of DRL to uncertain environments. Therefore, this method has significant theoretical and practical value for promoting the intelligent transformation of the energy system.

Keywords: integrated energy system of the park; deep reinforcement learning; deep Q-network; deep deterministic policy gradient; energy management (search for similar items in EconPapers)
JEL-codes: Q Q0 Q4 Q40 Q41 Q42 Q43 Q47 Q48 Q49 (search for similar items in EconPapers)
Date: 2025
References: Add references at CitEc
Citations:

Downloads: (external link)
https://www.mdpi.com/1996-1073/18/19/5172/pdf (application/pdf)
https://www.mdpi.com/1996-1073/18/19/5172/ (text/html)

Related works:
This item may be available elsewhere in EconPapers: Search for items with the same title.

Export reference: BibTeX RIS (EndNote, ProCite, RefMan) HTML/Text

Persistent link: https://EconPapers.repec.org/RePEc:gam:jeners:v:18:y:2025:i:19:p:5172-:d:1760704

Access Statistics for this article

Energies is currently edited by Ms. Cassie Shen

More articles in Energies from MDPI
Bibliographic data for series maintained by MDPI Indexing Manager ().

 
Page updated 2025-09-30
Handle: RePEc:gam:jeners:v:18:y:2025:i:19:p:5172-:d:1760704