Electric Power ›› 2023, Vol. 56 ›› Issue (3): 36-46.DOI: 10.11930/j.issn.1004-9649.202206099

• New Utilization of Renewable Energy and Its Key Technologies Under the Dual-Carbon Target • Previous Articles     Next Articles

Self-evolving Power Smooth Control Method for Offshore Wind Power Cluster Based On Deep Reinforcement Learning

SONG Weiye1, LIU Lingyue2, YAN Jie1, WANG Hangyu1, HE Shukai1, HAN Shuang1, WANG Minghui3, LIU Yongqian1   

  1. 1. State Key Laboratory of Alternate Electrical Power System with Renewable Energy Sources, North China Electric Power University, Beijing 102206, China;
    2. Nanjing Institute of Technology, Nanjing 211167, China;
    3. Jiangsu Goldwind Software Technology Co., Ltd., Wuxi 214000, China
  • Received:2022-06-13 Revised:2022-12-18 Accepted:2022-09-11 Online:2023-03-23 Published:2023-03-28
  • Supported by:
    This work is supported by the National Key Research and Development Program of China (No.2019YFE0104800) , the Young Elite Scientists Sponsorship Program by CAST (No.2019QNRC001) and Offshore Wind Power and Intelligent Energy System Project (Phase I) of Huaneng Group (No.HNKJ20-H88).

Abstract: The offshore wind speed has a high spatial-temporal correlation, which aggravates the power fluctuation of the whole wind farm and poses significant challenges to the power system, especially when large-scale offshore wind power is integrated. Smoothing control of large-scale offshore wind power clusters is a key solution to mitigating the above problems. However, most existing methods are inefficient and difficult to support higher frequency control and are susceptible to wind power forecast errors and the deviation of actual action from the optimal control command. Therefore, this paper proposes a new control framework for “offline-training, online-optimization and self-evolution”, and establishes a deep-reinforcement-learning-based model for the smoothing control of the active power of offshore wind power clusters. Firstly, a short-term revenue function for cluster power smoothing control is proposed to solve the optimal command based on the Markov decision process model. Secondly, a long-term revenue policy function for power policy calibration is proposed to effectively correct the control deviation according to the historical feedback data. Finally, a deep neural network model is established for the mapping between the state of the agent, the control benefit and the control decision to realize the training and solution of the agent based on the deep deterministic policy gradient algorithm. The results show that the proposed method can reduce the power fluctuation by 20% and control the power loss within 5% under the given wind condition of 7.5 m/s average wind speed.

Key words: wind power cluster, active power control, deep reinforcement learning, fluctuation smoothing, control deviation correction