共 52 条
Privacy-Preserving Probabilistic Voltage Forecasting in Local Energy Communities
被引:22
作者:
Toubeau, Jean-Francois
[1
]
Teng, Fei
[2
]
Morstyn, Thomas
[3
]
Von Krannichfeldt, Leandro
[4
]
Wang, Yi
[4
]
机构:
[1] Univ Mons, Power Syst & Markets Res Grp, Natl Fund Sci Res, B-7000 Mons, Belgium
[2] Imperial Coll London, Dept Elect & Elect Engn, London SW7 2AZ, England
[3] Univ Edinburgh, Sch Engn Sci, Edinburgh EH9 3JW, Midlothian, Scotland
[4] Univ Hong Kong, Dept Elect & Elect Engn, Hong Kong, Peoples R China
基金:
英国工程与自然科学研究理事会;
关键词:
Predictive models;
Data models;
Forecasting;
Voltage;
Training;
Privacy;
Probabilistic logic;
Differential privacy;
deep learning;
federated learning;
heterogeneous data;
voltage forecasting;
INFERENCE;
D O I:
10.1109/TSG.2022.3187557
中图分类号:
TM [电工技术];
TN [电子技术、通信技术];
学科分类号:
0808 ;
0809 ;
摘要:
This paper presents a new privacy-preserving framework for the short-term (multi-horizon) probabilistic forecasting of nodal voltages in local energy communities. This task is indeed becoming increasingly important for cost-effectively managing network constraints in the context of the massive integration of distributed energy resources. However, traditional forecasting tasks are carried out centrally, by gathering raw data of end-users in a single database that exposes their private information. To avoid such privacy issues, this work relies on a distributed learning scheme, known as federated learning wherein individuals' data are kept decentralized. The learning procedure is then augmented with differential privacy, which offers formal guarantees that the trained model cannot be reversed-engineered to infer sensitive local information. Moreover, the problem is framed using cross-series learning, which allows to smoothly integrate any new client joining the community (i.e., cold-start forecasting) without being plagued by data scarcity. Outcomes show that the proposed approach achieves improved performance compared to non-collaborative (locally trained) models, and is able to reach a trade-off between privacy and performance for different architectures of deep learning networks.
引用
收藏
页码:798 / 809
页数:12
相关论文