中国电力 ›› 2023, Vol. 56 ›› Issue (3): 36-46.DOI: 10.11930/j.issn.1004-9649.202206099

• 双碳目标下可再生能源新型利用及关键技术 • 上一篇    下一篇

基于深度强化学习的海上风电集群自进化功率平滑控制方法

宋伟业1, 刘灵玥2, 阎洁1, 王航宇1, 何书凯1, 韩爽1, 王明辉3, 刘永前1   

  1. 1. 新能源电力系统国家重点实验室(华北电力大学),北京 102206;
    2. 南京工程学院, 江苏 南京 211167;
    3. 江苏金风软件技术有限公司,江苏 无锡 21400
  • 收稿日期:2022-06-13 修回日期:2022-12-18 出版日期:2023-03-28 发布日期:2023-03-28
  • 作者简介:宋伟业(1999-),男,硕士研究生,从事风电场控制、强化学习算法研究,E-mail:120222111005@ncepu.edu.cn;阎洁(1987-),女,通信作者,博士,副教授,从事新能源场站功率预测与控制研究,E-mail:yanjie@ncepu.edu.cn;王航宇(1996-),男,硕士研究生,从事风电场控制、强化学习算法研究,E-mail:hangyuw@ncepu.edu.cn
  • 基金资助:
    国家重点研发计划资助项目(2019YFE0104800);中国科协青年人才托举工程项目(2019QNRC001);华能集团总部科技项目“海上风电与智慧能源系统科技专项(一期)”(HNKJ20-H88)。

Self-evolving Power Smooth Control Method for Offshore Wind Power Cluster Based On Deep Reinforcement Learning

SONG Weiye1, LIU Lingyue2, YAN Jie1, WANG Hangyu1, HE Shukai1, HAN Shuang1, WANG Minghui3, LIU Yongqian1   

  1. 1. State Key Laboratory of Alternate Electrical Power System with Renewable Energy Sources, North China Electric Power University, Beijing 102206, China;
    2. Nanjing Institute of Technology, Nanjing 211167, China;
    3. Jiangsu Goldwind Software Technology Co., Ltd., Wuxi 214000, China
  • Received:2022-06-13 Revised:2022-12-18 Online:2023-03-28 Published:2023-03-28
  • Supported by:
    This work is supported by the National Key Research and Development Program of China (No.2019YFE0104800) , the Young Elite Scientists Sponsorship Program by CAST (No.2019QNRC001) and Offshore Wind Power and Intelligent Energy System Project (Phase I) of Huaneng Group (No.HNKJ20-H88).

摘要: 海上风电集群的风速时空相关性强,加剧了整体有功输出的波动幅度,大规模并网对电力系统的影响更为突出,海上风电集群有功输出平滑控制是解决上述问题的关键手段。传统方法的优化效率低、难以支撑高频率控制,且对预测误差、执行偏差灵敏度过高。因此,提出了“策略离线训练、在线快速寻优、控制效果自进化”的控制架构,建立了深度强化学习的海上风电集群有功输出平滑控制模型。首先,提出了面向集群功率平滑控制的短期收益函数,基于马尔科夫决策过程模型求解最优指令;其次,提出了面向功率策略校准的长期收益Policy函数,根据历史反馈数据有效矫正控制偏差;最后,建立了智能体状态、控制收益和控制决策之间映射的深度神经网络模型,实现基于深度确定性策略梯度算法的智能体训练与求解。算例结果表明:在平均风速为7.5m/s的给定风况下,所提方法能够降低功率波动幅度达20%,同时将发电量损失控制在5%以内。

关键词: 风电集群, 有功功率控制, 深度强化学习, 波动平抑, 控制偏差修正

Abstract: The offshore wind speed has a high spatial-temporal correlation, which aggravates the power fluctuation of the whole wind farm and poses significant challenges to the power system, especially when large-scale offshore wind power is integrated. Smoothing control of large-scale offshore wind power clusters is a key solution to mitigating the above problems. However, most existing methods are inefficient and difficult to support higher frequency control and are susceptible to wind power forecast errors and the deviation of actual action from the optimal control command. Therefore, this paper proposes a new control framework for “offline-training, online-optimization and self-evolution”, and establishes a deep-reinforcement-learning-based model for the smoothing control of the active power of offshore wind power clusters. Firstly, a short-term revenue function for cluster power smoothing control is proposed to solve the optimal command based on the Markov decision process model. Secondly, a long-term revenue policy function for power policy calibration is proposed to effectively correct the control deviation according to the historical feedback data. Finally, a deep neural network model is established for the mapping between the state of the agent, the control benefit and the control decision to realize the training and solution of the agent based on the deep deterministic policy gradient algorithm. The results show that the proposed method can reduce the power fluctuation by 20% and control the power loss within 5% under the given wind condition of 7.5 m/s average wind speed.

Key words: wind power cluster, active power control, deep reinforcement learning, fluctuation smoothing, control deviation correction