FedLbs: Federated Learning Loss-Based Swapping Approach for Energy Building’s Load Forecasting
Conférence : Communications avec actes dans un congrès international
Federated Learning (FL) is rapidly growing in popularity as a decentralized approach and is being adopted in smart building systems and energy forecasting without accessing sensitive data. Specifically, clients train their models using their own data. After that, only their model parameters are sent to the central server, which aggregates them by averaging the weights and then sends back the newly formed model to each client. However, challenges arise when dealing with heterogeneous multivariate time-series data with different distributions. This leads to higher-performing clients contributing to the global update more than the others, and slower convergence where the global model takes more time to generalize across the clients. In this paper, we propose an enhanced aggregation approach, where the server sorts clients’ models based on their local training losses before swapping them all consecutively according to the best and worst-performing ones. Our proposed approach is applied to a smart building dataset and compared with two other FL approaches from the literature. Our simulation results demonstrate improved forecasting precision for each client and faster convergence. Moreover, we optimized the global model’s evaluation error scores and overall loss, reduced the communication rounds required for convergence, and ensured less bias and more fairness between clients during each training cycle.