Next Article in Journal
SR-FEINR: Continuous Remote Sensing Image Super-Resolution Using Feature-Enhanced Implicit Neural Representation
Previous Article in Journal
A Model for Cognitive Personalization of Microtask Design
Previous Article in Special Issue
A Two-Stage Multi-Agent EV Charging Coordination Scheme for Maximizing Grid Performance and Customer Satisfaction
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

FedBranched: Leveraging Federated Learning for Anomaly-Aware Load Forecasting in Energy Networks

1
James Watt School of Engineering, University of Glasgow, Glasgow G12 8QQ, UK
2
Department of Electrical Engineering, University of Engineering and Technology, Lahore-Faisalabad Campus, Faisalabad 38000, Pakistan
3
Thomas Johann Seebeck Department of Electronics, Tallinn University of Technology, 19086 Tallinn, Estonia
*
Author to whom correspondence should be addressed.
Sensors 2023, 23(7), 3570; https://doi.org/10.3390/s23073570
Submission received: 22 February 2023 / Revised: 16 March 2023 / Accepted: 27 March 2023 / Published: 29 March 2023
(This article belongs to the Special Issue Machine Learning Techniques for Energy Efficient IoT Networks)

Abstract

:
Increased demand for fast edge computation and privacy concerns have shifted researchers’ focus towards a type of distributed learning known as federated learning (FL). Recently, much research has been carried out on FL; however, a major challenge is the need to tackle the high diversity in different clients. Our research shows that using highly diverse data sets in FL can lead to low accuracy of some local models, which can be categorised as anomalous behaviour. In this paper, we present FedBranched, a clustering-based framework that uses probabilistic methods to create branches of clients and assigns their respective global models. Branching is performed using hidden Markov model clustering (HMM), and a round of branching depends on the diversity of the data. Clustering is performed on Euclidean distances of mean absolute percentage errors (MAPE) obtained from each client at the end of pre-defined communication rounds. The proposed framework was implemented on substation-level energy data with nine clients for short-term load forecasting using an artificial neural network (ANN). FedBranched took two clustering rounds and resulted in two different branches having individual global models. The results show a substantial increase in the average MAPE of all clients; the biggest improvement of 11.36% was observed in one client.

1. Introduction

With the advent of the Internet of Things (IoT), a significant rise in connected devices and sensors empowered by the 5G communication network generated a large amount of data. The overwhelming availability of data captured by edge devices, coupled with advanced machine learning (ML), inspired many researchers to process data on a massive scale for real-world applications [1]. However, the traditional cloud-centric ML training approach uses centralised data, which requires a large amount of data to be transferred from the end devices to a third-party central server [2]. The data generated by edge devices are privacy sensitive and highly protected under the law [3]. Furthermore, raw data transfer also burdens the communication networks and incurs latency caused by long propagation delays, which is not acceptable in time-sensitive applications [4].
Therefore, to overcome these challenges, a distributed learning approach called federated learning (FL) was introduced. FL aims to train ML models collaboratively, orchestrated by a centralised server [5]. FL is collaborative and continual learning in which a global model is updated by aggregating the local models’ parameters using training on edge devices. In centralised model training, it is assumed that the data are independent and identically distributed (IID). However, due to the decentralised nature of data, this assumption is not applicable in the FL setting. Therefore, federated averaging (FedAVG) is one of the simplest and most commonly used techniques for model aggregation for non-IID data, whereby the weighted average of model parameters is performed on the central server [6]. This model training approach offers privacy by design, because no data are shared.
Short-term load forecasting (STLF) is an essential component of smart grid systems because it allows utilities and grid managers to predict electricity consumption for the coming hours or days. Accurate load forecasting enables better planning and management of resources, which can aid in cost reduction, energy optimization, and grid resilience [7]. Statistical time series techniques such as linear or non-parametric regression, exponential smoothing, auto-regressive moving average, and auto-regressive integrated moving average (ARIMA) are present in the literature [8,9]. However, with the rise of big data and artificial intelligence (AI), more advanced machine learning (ML) algorithms, such as deep neural networks, have been developed to map complicated and non-linear behaviour in historical load profiles. Unfortunately, DL requires the transmission of massive amounts of historical data to a centralised location for model training [10]. Although centralised model training has achieved promising results, it also presents multiple challenges, including data privacy, transmission costs, and secure island data access. Furthermore, the energy market is fiercely competitive, and utility companies are reluctant to share data with their competitors. FL has the potential to solve the challenges mentioned above and train a robust forecasting model without sharing any data.
Despite various advantages, FL has to deal with challenges such as device heterogeneity, privacy, data imbalances, secure aggregation, data diversity, and anomalous clients. For instance, various devices participating in the training process may have diverse computation and storage power, leading to different convergence times. Privacy is often a significant concern in data-driven applications, which FL can solve. However, communicating model updates throughout training can reveal sensitive information to a third party or the central server. This issue is often overcome using homomorphic encryption and differential privacy [11]. Because the server does not have access to the clients’ data, there is the possibility that a client might not be sending proper updates to the server. This problem and its resolution have been discussed in detail in [12]. Similarly, data diversity is one of the biggest challenges in the FL environment. For instance, the model’s accuracy is significantly reduced, by 55%, when it is trained on highly skewed data [13]. Therefore, one possible solution is to share a small amount of data to perform statistical analysis for clients’ clustering on the server. However, data sharing is not feasible due to privacy concerns. Another possible method to address data heterogeneity is clustering the users based on the statistical properties of the model’s parameters on the server. The idea is to obtain multiple global models to improve the system’s performance. In this work, we propose a hidden Markov model (HMM)-based clustering approach, creating multiple global models. The key idea is to monitor each client’s mean absolute percentage error (MAPE) and compute the Euclidean distance of MAPE for each client. The branch on each client represents a separate global model for clients with similar data.

1.1. Related Work

STLF involves predicting the electricity demand of a given area over a period of a few hours or days. This is crucial for energy suppliers because it allows them to optimise energy production resources to eliminate power outages and reduce costs. This has traditionally been achieved by collecting data from smart meters to train a centralised model. However, this entails exchanging privacy-sensitive information with a central server, posing serious privacy concerns. Although FL has shown promising results in various applications, there are very limited studies to date on its use for smart grid applications, especially STLF. For instance, the authors of [14] proposed a deep neural network (DNN)-based STLF model using smart meter data. The proposed method uses differential privacy for security, and performance is evaluated by computing the mean absolute percentage error (MAPE). An LSTM-based neural network was proposed for STLF in an agile workspace at the University of Glasgow, UK, with the help of the Persuasive Energy Conscious Network (PECN) [15]. The authors of [9] propose residential energy forecasting using FL and edge computing to ensure the privacy of energy consumption data. The effectiveness of the proposed model is evaluated through experiments using different configurations with varying levels of privacy and compared against a traditional centralised approach. The results revealed that increasing privacy levels typically came at the cost of decreased accuracy in the overall prediction, as demonstrated by the scenarios with different edge-computing configurations. Similarly, a non-clustering-based approach to train long-short-term memory (LSTM) networks for STLF is proposed in [16]. In this study, individual households are used as federated clients, and the results reveal that FL is not suitable for individual households as they have very unusual energy consumption patterns. Additionally, STLF is a complex task that involves many variables, and the effectiveness of FL for this task may depend on the specific characteristics of the data and the energy system being modelled.
In FL, the performance of the global model is highly dependent on the quality of data on each client. Diversity in data significantly impacts the overall performance of the system. The commonly used FedAVG algorithm for model aggregation does not compensate for the divergence in model weights caused by non-IID data. For instance, the authors of [13] analysed the performance of a convolutional neural network (CNN) using non-IID data. The results show a significant divergence between local model weights’ means and standard deviations, resulting in performance degradation. To overcome this challenge, a small subset of global data is created and shared among clients for clustering to improve the model’s accuracy. The authors of [17] proposed an auto-encoder-based clustering approach to group patients’ electronic records into several communities. For each community, a separate model is trained simultaneously based on the similarity of the data, resulting in improved results. A hierarchical clustering approach to deal with non-IID data in FL is proposed in [18]. The idea is to capitalise on the similarity of local updates to create a specialised global model for each subset of clients. This approach not only improves the model’s overall performance, but also reduces the convergence time.
Similarly, the authors of [19] proposed a client clustering approach based on the similarities of gradients computed in local updates. This technique improves the accuracy and communication cost of the system. In [20], a soft clustering algorithm is proposed, using overlapping clusters to cater to the complex nature of the data. The information on overlapping clients can be utilised to categorise similar clients more efficiently to enhance the system’s performance. The authors of [21] proposed FedSim, a clustering approach based on the pairwise similarity of the local gradients. The proposed FedSim algorithm decomposes the training process into local and global steps, aggregating clients with similar gradients to ensure better convergence. The results are compared with those of traditional FedAVG, and the findings of this study confirm that exploiting latent inter-client similarities significantly improves performance.

1.2. Motivation and Contributions

With the existing FL mechanism, a single global model is obtained using iterative local updates on the server. However, the data from edge devices are highly diverse, which causes a significant divergence in model parameters, resulting in performance degradation as analysed in [13]. Therefore, multiple global models are the way forward to deal with the diversity of client data. One possible solution is using a global data set for client clustering, which poses serious privacy concerns. Furthermore, federated personalisation is another way of dealing with diverse data, whereby the optimal global model can be aggregated either layer-wise or through model architecture to obtain a personalised model. This personalised training increases the model’s convergence time, as each client needs training for the optimal local model [22]. Another commonly used technique to ensure privacy is based on measuring the similarity of model parameters such as gradient and mean or variance of layerwise weights. The similarity measure is an efficient method of client clustering to deal with data diversity. However, to ensure the privacy of the global aggregation, differential privacy, whereby random noise is added to local weights, is used. With differential privacy, keeping track of similarity measures is challenging, as noise levels will affect model parameters differently [11]. Therefore, to deal with this issue, this work proposes an HMM clustering approach which uses the local loss function MAPE as a metric for measuring the similarity of clients. The aim is to use the Euclidean MAPE distance of each client to develop a branch which represents multiple global models using a subset of local clients. The key contributions of this work are highlighted below:
  • We propose a zero-knowledge-based client clustering approach using HMM, which uses the Euclidean distances of loss functions. In each communication round, the clients share the model parameter and value of the loss function, in our case MAPE, to measure the similarity among different clients.
  • The proposed FedBranched framework deals with diverse client data in the FL environment to make multiple global models based on the level of data diversity. Furthermore, there is no predefined limit for the number of clusters, which makes this approach more flexible for highly diverse data.
  • The use of FedBranched guarantees convergence of the loss function.
  • We propose an ANN and selected features for short-term energy forecasting.
  • We perform an analysis of energy consumption before and after applying FedBranched.

1.3. Paper Organisation

The rest of the paper is organised as follows: Section 2 discusses the proposed framework, and Section 3 explains the simulation setup. Section 4 provides a discussion of the experimental setup and results. The conclusions and future research directions are given in Section 5.

2. FedBranched

FedBranched is a zero-knowledge framework that uses HMM clustering to create branches of clients based on the Euclidean distances of MAPE of each client, and each branch is assigned a new global model. The HMM is a generative probabilistic model in which a series of internal hidden states Z generates a series of observable X variables. Direct observation of the hidden states is not possible. It is assumed that the transitions between hidden states take the shape of a (first-order) Markov chain. The use of HMM in the clustering mechanism is explained in detail in [23].
There can be multiple stages of clustering; however, clustering will only create two branches at any stage. The idea behind restricting the number of branches to two is to reduce the total number of required global models. The maximum number of branches can be n/2, where n is the total number of clients.
At the beginning of FedBranched, a generic ML model is sent to all the clients. When data stored locally are passed through this ML model, the newly generated ML is then called the local model. All the devices send local models and a loss function to the server, and the server aggregates them to make a new model, known as a global model, M. We chose MAPE as the loss function, which can be calculated as:
MAPE = 1 n n t = 1 A t F t A t
where A t is the actual value, F t is the forecasted value, and n is the total number of iterations. It is worth mentioning that the MAPE is independent of the system capacity and the unit of measurement. It may be the only error metric that can be used to compare forecasting performance between various utilities [24].
The generation of the global model completes one communication round. After a pre-decided number of communication rounds, the server analyses the MAPE of all clients and identifies non-convergent clients. If the MAPE of all clients does not converge, then the server creates two groups of clients, known as Branch 1 and Branch 2. To make two new branches, the server calculates each client’s Euclidean distances of MAPE and places it in a n n matrix. Then, HMM clustering on the sum of each row is used to create two branches. The server now sends a generic ML to both branches, and local branch ML models are sent back to be aggregated into new mini-global models, M 1 and M 2 . At this stage, there are two global models. This completes one clustering round. Now, the server again checks the convergence of MAPE of all clients in both branches. If the MAPE of any client in any of the branches does not converge, then another round of clustering is required. Branch 1 is divided further into Branches 3 and 4, while Branch 2 is divided into Branches 5 and 6. This only happens if the MAPE of all clients in both branches, 1 and 2, could not converge. If any branches from 1 or 2 converge, then newly gendered branches from non-converging branches will be merged individually with the previously converged branch.
Furthermore, the new global model is decided for that branch. For example: the MAPE of all clients in Branch 1 did not converge; however, the MAPE of all clients in Branch 2 has converged. In this case, the client in Branch 1 is sliced into Branches 3 and 4. Now, Branches 3 and 4 are merged with Branch 1 individually, and new mini-global models M 1 , 3 and M 1 , 4 are created for these branches. If the MAPE of all clients converges in these branches, we will still have two global models. If a MAPE does not converge, we discard M 1 , 3 and M 1 , 4 and proceed to make global models from Branches 3 and 4, M 3 and M 4 . Moreover, the branching process continues until all the MAPE of all the clients in all the branches have converged. This process is explained graphically in Figure 1. Here, N represents the total number of global models at any stage.

3. Simulation Setup

To test the results of FedBranched and compare them with those of Vanilla FL, we created an FL environment with nine clients. We selected a real-world energy data set from PJM Interconnection, LLC [25]. In this data set, each column represents energy utilised by each substation. A data set sample is plotted in Figure 2. Each client has 13,896 samples. It can be seen that it is a very diverse data set, and the clients do not resemble one another. The letter-value plot in Figure 3 can further confirm the diversity. It can be visualised that Client 1 and Client 7 have the highest and lowest mean, respectively. Furthermore, it can also be noticed that Client 2 has the highest number of outliers. This data set is used for short-term load forecasting (STLF) with the help of five features. The included features were last hour value, last day value, last week value, last 24 h average, and last week average. An artificial neural network (ANN) with three dense layers was designed for STLF. The first layer had one hundred neurons, the second layer had fifty, and the last layer had one neuron. All the layers had relu activation functions and Adam optimiser with mean square error as the loss function. We used a 70/30 split for training and test data. We ran FL for 30 communication rounds; each client had 15 local epochs and a batch size of 300. We selected FedAVG [26] as the aggregation method at the server.

4. Experiment and Results

The results of traditional FL are plotted in Figure 4, and here, the MAPE of each client from local models with each communication round is reported. It can be seen that the MAPE of all the clients converges at the eighth communication round except for Client 6 and Client 7. The best and worst MAPE, of 2.51% and 15.71%, were produced by Clients 2 and 7, respectively, after 30 communication rounds. It is evident that a single global model is insufficient, and that Clients 6 and 7 should be treated with a separate global model. The original and predicted data of each client is presented in Figure 5. Here, consumed energy is plotted for the sample number. All clients can predict comparatively better than Clients 6 and 7.

4.1. Clustering Round 1

Because the MAPE of all clients did not converge, HMM clustering was applied to divide the clients into two branches. The results of clustering are presented in Figure 6. Here, Clients 1, 2, 3, 4, 5, 8, and 9 were clustered together (Branch 1), while Clients 6 and 7 were clustered together (Branch 2). Next, two different mini-global models were created for each branch. The MAPE of each client of Branches 1 and 2 are plotted in Figure 7a,b, respectively. It can be seen that, with the MAPE obtained during model aggregation in 30 communication rounds in both branches, the MAPE of all clients except Client 3 from Branch 1 converged. The original and predicted curves are plotted in Figure 8 and Figure 9. In Branch 1, the lowest and highest MAPE, of 2.5% and 18.96%, were obtained from Clients 8 and 3, respectively. In Branch 2, MAPE of 2.6% and 4.3% was obtained from Clients 6 and 7, respectively. At this stage, two mini-global models, M 1 and M 2 , were created for Branches 1 and 2, respectively. However, because the MAPE of Branch 1 did not converge during model aggregation, we can discard M 1 and apply HMM for further clustering. It is worth noticing that the MAPE of Client 3 converged when all the clients were grouped during model aggregation, as seen in Figure 4, which means Client 3 performs better with clients in the other branch. Furthermore, it can be observed that Branch 1 required two global models.

4.2. Clustering Round 2

We applied HMM clustering to Branch 1, and the results are plotted in Figure 10. Clients 1, 2, 4, 5, 8, 9 are placed in Branch 3, and Client 3 in Branch 4. According to the FedBranched framework, before making new global models for Branches 3 and 4, both must be merged with the previously converging branch (Branch 2). When Branch 4 was converged with Branch 2 (Clients 3, 6, and 7) a new global model, M 2 , 4 , was generated. The MAPE of all clients in Branches 3 and 4 is plotted with respective communication rounds in Figure 11a,b. It can be seen that the MAPE of all clients converged. Because merging Branch 2 and 4 produced converging MAPE during the model aggregation, there is no need to merge Branch 3 with Branch 2 because we cannot place the same clients in two different branches. A new mini-global model, M 3 , was created for Branch 3, having Clients 1, 2, 4, 5, 8, and 9. The MAPE of all clients during model aggregation are plotted in Figure 11b. It can be seen that the MAPE of all the clients converged during model aggregation.

4.3. Discussion

After running FedBranched for two clustering rounds, only two mini-global models, M 2 , 4 and M 3 , were finalised. The original and predicted curves from both branches are plotted in Figure 12 and Figure 13. The results of traditional FL and FedBranched are summarised in Table 1. The highest percentage improvement, of 11.36%, is observed for Client 7, and the lowest is −0.32%, for Client 4. The clustering mechanism using FedBranched on the given data is graphically elaborated in Figure 14. It can be seen that Clients 3, 4, and 7 were clustered together while Clients 1, 2, 3, 4, 5, 8 and 9 were put into another cluster. For both clusters, individual global models were assigned. Our results suggest that FedBranched can successfully improve the forecasting accuracy of highly diverse data. The average MAPE of all clients using Vanilla FL was 5.172%; after applying FedBranched, it dropped to 2.83%.

5. Design Insights and Analysis

In this paper, the authors present a framework that requires zero knowledge of the data set and uses a probability-based clustering approach for heterogeneous data in FL. This framework ensures the convergence of the loss function.
  • Key features: FedBranched is capable of clustering clients without looking at the user’s data. Clustering is performed on the sum of Euclidean distances of a loss function using HMM. Here, a multi-stage clustering mechanism is adopted to minimize the required clusters and global models. At any stage, there can be only two clusters. The purpose of restricting the number of clusters is to have control and restrict the number of clusters to a minimum.
  • Scope: FedBranched uses loss functions rather than weights of ML models; it can work perfectly with differential privacy [27] and homomorphic encryption [28].
  • Limitations: Because it is a clustering-based approach, it might not work well when the number of clients is less than five.
  • Drawbacks: FedBranched uses a multistage clustering approach depending upon the diversity of data. At each stage, new branches are created and their respective global models are decided after running FL for pre-decided communication rounds. This will increase the computation power and time required to achieve an optimal number of global models that can satisfy the user’s requirements. In this paper, Vanilla FL was carried out for 30 communication rounds, whereas FedBranched was run for 150 (30*5) communication rounds, as shown in Figure 14.
  • Energy consumption: Energy consumption, E c o m , in the training process of a global model depends on many factors, such as energy consumed per kb when local and global data are shared between clients and server, time of data transfer, type of IoT device used, the communication channel used, computation time on each device, etc. Here, E c o m is described based on the work of [29]:
    E c o m = R [ ( α t ) + ( β D ) ]
    where R represents number of communication rounds, α represents energy per second, t is for time, β represents energy consumed per kb, and D represents amount of data transfer. In our experiments, each global or local model is of 38 kb and β is taken as 0.015 kWh/GB from [30], while α is assumed as 0.0001 kWh/sec; then, baseline E c o m , when t is 3.39 min, is equal to 27 W. However, using FedBranched, E c o m reached 136 W.

6. Conclusions and Future Work

The use of diverse data sets in FL can lead to high forecasting error in some local models, which can be categorised as an anomaly. In this paper, the authors have presented FedBranched, a zero-knowledge clustering-based framework to tackle diverse data sets in FL. Clustering is performed using HMM clustering on the Euclidean distance of MAPE obtained after a pre-decided number of communication rounds. This framework divides clients into different branches and assigns their respective global models. Clustering is performed in multiple rounds depending upon the diversity of data, and there can be a maximum of n / 2 clusters where n is the total number of clients; however, at any stage, there can only be two branches. The developed framework was applied to a highly diverse, real-life energy data set obtained at the substation level, having nine clients, for STLF using ANN. FL was carried out for thirty communication rounds, and the ML model was trained for fifteen epochs in each round. The initial results of FL showed that one global model is not enough for nine clients. FedBranched took two clustering rounds and ended up dividing nine clients into two clusters and two global models. The use of FedBranched resulted in a noticeable increase in MAPE. The highest improvement of 11.6% was observed for Client 6. However, energy consumption analysis shows that during the training process, the global model Vanilla FL consumed only 27 W, while FedBranched used 136 W. The authors are currently working on an extension of this work, which aims to design an efficient framework that uses all of the possible combinations of clients and transfer FL to make a global model that provides the lowest loss function. This will remove the need for multiple global models.

Author Contributions

Conceptualization: H.U.M., A.R.K., A.Z.; Data curation: H.U.M., A.R.K.; Formal Analysis: H.U.M., D.F., M.M.A.; Methodology: H.U.M., A.R.K., M.A.I., A.Z.; Resources: D.F., M.M.A., M.A.; Software: H.U.M., A.R.K.; Original Draft Preparation: H.U.M., A.R.K.; Review and Editing: H.U.M., A.R.K., D.F., M.M.A., M.A., M.A.I., A.Z.; Supervision: A.Z. All authors have read and agreed to the published version of the manuscript.

Funding

This project has received funding from the European Union’s Horizon 2020 Research programme under grant agreement No. 101058505.

Data Availability Statement

The data set used is available at: www.kaggle.com/datasets/robikscube/hourly-energy-consumption (accessed on 8 August 2022).

Acknowledgments

This project has received funding from the European Union’s Horizon 2020 Research programme under grant agreement No. 101058505. This research has also been supported in part by the Internet of Intelligent Things project of Estonian IT Academy program and Estonian Research Council under Grant PUT-PRG424.

Conflicts of Interest

The authors declare no conflict of interest.

Abbreviations

The following abbreviations are used in this manuscript:
FLFederated learning
MLMachine learning
nTotal number of clients
NTotal number of global models
MGlobal model
ANNArtificial neural network
STLFShort-term load forecasting

References

  1. Lim, W.Y.B.; Luong, N.C.; Hoang, D.T.; Jiao, Y.; Liang, Y.C.; Yang, Q.; Niyato, D.; Miao, C. Federated learning in mobile edge networks: A comprehensive survey. IEEE Commun. Surv. Tutor. 2020, 22, 2031–2063. [Google Scholar]
  2. Li, P.; Li, J.; Huang, Z.; Li, T.; Gao, C.Z.; Yiu, S.M.; Chen, K. Multi-key privacy-preserving deep learning in cloud computing. Future Gener. Comput. Syst. 2017, 74, 76–85. [Google Scholar] [CrossRef]
  3. Custers, B.; Sears, A.M.; Dechesne, F.; Georgieva, I.; Tani, T.; van der Hof, S. EU Personal Data Protection in Policy and Practice; Springer: Cham, Switzerland, 2019. [Google Scholar]
  4. Yang, K.; Jiang, T.; Shi, Y.; Ding, Z. yang2020federated. IEEE Trans. Wirel. Commun. 2020, 19, 2022–2035. [Google Scholar] [CrossRef]
  5. McMahan, H.B.; Moore, E.; Ramage, D.; y Arcas, B.A. Federated learning of deep networks using model averaging. arXiv 2016, arXiv:1602.05629. [Google Scholar]
  6. Konečnỳ, J.; McMahan, H.B.; Yu, F.X.; Richtárik, P.; Suresh, A.T.; Bacon, D. Federated learning: Strategies for improving communication efficiency. arXiv 2016, arXiv:1610.05492. [Google Scholar]
  7. Moradzadeh, A.; Mohammadpourfard, M.; Konstantinou, C.; Genc, I.; Kim, T.; Mohammadi-Ivatloo, B. Electric load forecasting under False Data Injection Attacks using deep learning. Energy Rep. 2022, 8, 9933–9945. [Google Scholar] [CrossRef]
  8. Khan, A.R.; Mahmood, A.; Safdar, A.; Khan, Z.A.; Khan, N.A. Load forecasting, dynamic pricing and DSM in smart grid: A review. Renew. Sustain. Energy Rev. 2016, 54, 1311–1322. [Google Scholar] [CrossRef]
  9. Petrangeli, E.; Tonellotto, N.; Vallati, C. Performance Evaluation of Federated Learning for Residential Energy Forecasting. IoT 2022, 3, 381–397. [Google Scholar] [CrossRef]
  10. Gholizadeh, N.; Musilek, P. Federated learning with hyperparameter-based clustering for electrical load forecasting. IoT 2022, 17, 100470. [Google Scholar] [CrossRef]
  11. Wei, K.; Li, J.; Ding, M.; Ma, C.; Yang, H.H.; Farokhi, F.; Jin, S.; Quek, T.Q.; Poor, H.V. Federated learning with differential privacy: Algorithms and performance analysis. IEEE Trans. Inf. Forensics Secur. 2020, 15, 3454–3469. [Google Scholar] [CrossRef]
  12. Manzoor, H.U.; Khan, M.S.; Khan, A.R.; Ayaz, F.; Flynn, D.; Imran, M.A.; Zoha, A. FedClamp: An Algorithm for Identification of Anomalous Client in Federated Learning. In Proceedings of the 2022 29th IEEE International Conference on Electronics, Circuits and Systems (ICECS), Glasgow, UK, 24–26 October 2022; IEEE: Piscataway, NJ, USA, 2022; pp. 1–4. [Google Scholar]
  13. Zhao, Y.; Li, M.; Lai, L.; Suda, N.; Civin, D.; Chandra, V. Federated learning with non-iid data. arXiv 2018, arXiv:1806.00582. [Google Scholar] [CrossRef]
  14. Savi, M.; Olivadese, F. Short-term energy consumption forecasting at the edge: A federated learning approach. IEEE Access 2021, 9, 95949–95969. [Google Scholar] [CrossRef]
  15. Manzoor, H.U.; Khan, A.R.; Al-Quraan, M.; Mohjazi, L.; Taha, A.; Abbas, H.; Hussain, S.; Imran, M.A.; Zoha, A. Energy Management in an Agile Workspace using AI-driven Forecasting and Anomaly Detection. In Proceedings of the 2022 4th Global Power, Energy and Communication Conference (GPECOM), Nevsehir, Turkey, 14–17 June 2022; IEEE: Piscatawy, NJ, USA, 2022; pp. 644–649. [Google Scholar]
  16. Taïk, A.; Cherkaoui, S. Electrical load forecasting using edge computing and federated learning. In Proceedings of the ICC 2020-2020 IEEE International Conference on Communications (ICC), Dublin, Ireland, 7–11 June 2020; IEEE: Piscataway, NJ, USA, 2020; pp. 1–6. [Google Scholar]
  17. Huang, L.; Shea, A.L.; Qian, H.; Masurkar, A.; Deng, H.; Liu, D. Patient clustering improves efficiency of federated machine learning to predict mortality and hospital stay time using distributed electronic medical records. J. Biomed. Inform. 2019, 99, 103291. [Google Scholar] [CrossRef] [PubMed]
  18. Briggs, C.; Fan, Z.; Andras, P. Federated learning with hierarchical clustering of local updates to improve training on non-IID data. In Proceedings of the 2020 International Joint Conference on Neural Networks (IJCNN), Glasgow, UK, 19–24 July 2020; IEEE: Piscataway, NJ, USA, 2020; pp. 1–9. [Google Scholar]
  19. Luo, Y.; Liu, X.; Xiu, J. Energy-efficient clustering to address data heterogeneity in federated learning. In Proceedings of the ICC 2021-IEEE International Conference on Communications, Montreal, QC, Canada, 14–23 June 2021; IEEE: Piscataway, NJ, USA, 2021; pp. 1–6. [Google Scholar]
  20. Li, C.; Li, G.; Varshney, P.K. Federated Learning With Soft Clustering. IEEE IoT J. 2021, 9, 7773–7782. [Google Scholar] [CrossRef]
  21. Sattler, F.; Müller, K.R.; Samek, W. Clustered federated learning: Model-agnostic distributed multitask optimization under privacy constraints. IEEE Trans. Neural Netw. Learn. Syst. 2020, 32, 3710–3722. [Google Scholar] [CrossRef] [PubMed]
  22. Ma, X.; Zhang, J.; Guo, S.; Xu, W. Layer-wised model aggregation for personalized federated learning. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 10092–10101. [Google Scholar]
  23. Ghassempour, S.; Girosi, F.; Maeder, A. Clustering multivariate time series using hidden Markov models. Int. J. Environ. Res. Public Health 2014, 11, 2741–2763. [Google Scholar] [CrossRef]
  24. Candelieri, A. Clustering and support vector regression for water demand forecasting and anomaly detection. Water 2017, 9, 224. [Google Scholar] [CrossRef]
  25. Mulla, R. Hourly Energy Consumption. Available online: https://www.kaggle.com/datasets/robikscube/hourly-energy-consumption (accessed on 8 August 2022).
  26. Zhou, Y.; Ye, Q.; Lv, J. Communication-efficient federated learning with compensated overlap-fedavg. IEEE Trans. Parallel Distrib. Syst. 2021, 33, 192–205. [Google Scholar] [CrossRef]
  27. Dwork, C. Differential privacy: A survey of results. In Proceedings of the International Conference on Theory and Applications of Models of Computation, Xi’an, China, 25–29 April 2008; Springer: Cham, Switzerland, 2008; pp. 1–19. [Google Scholar]
  28. Fontaine, C.; Galand, F. A survey of homomorphic encryption for nonspecialists. EURASIP J. Inf. Secur. 2007, 2007, 1–10. [Google Scholar] [CrossRef]
  29. Mian, A.N.; Shah, S.W.H.; Manzoor, S.; Said, A.; Heimerl, K.; Crowcroft, J. A value-added IoT service for cellular networks using federated learning. Comput. Netw. 2022, 213, 109094. [Google Scholar] [CrossRef]
  30. Aslan, J.; Mayers, K.; Koomey, J.G.; France, C. Electricity intensity of internet data transmission: Untangling the estimates. J. Ind. Ecol. 2018, 22, 785–798. [Google Scholar] [CrossRef]
Figure 1. Framework of FedBranched.
Figure 1. Framework of FedBranched.
Sensors 23 03570 g001
Figure 2. A sample of the data set.
Figure 2. A sample of the data set.
Sensors 23 03570 g002
Figure 3. Diversity visualization of the data set.
Figure 3. Diversity visualization of the data set.
Sensors 23 03570 g003
Figure 4. FL results with one global model after 30 communication rounds.
Figure 4. FL results with one global model after 30 communication rounds.
Sensors 23 03570 g004
Figure 5. Prediction results with one global model.
Figure 5. Prediction results with one global model.
Sensors 23 03570 g005
Figure 6. HMM clustering results for round 1. Here, red and green colours are used for different clusters.
Figure 6. HMM clustering results for round 1. Here, red and green colours are used for different clusters.
Sensors 23 03570 g006
Figure 7. First round of clustering. (a) FL Result of Branch 2. (b) FL Result of Branch 1.
Figure 7. First round of clustering. (a) FL Result of Branch 2. (b) FL Result of Branch 1.
Sensors 23 03570 g007
Figure 8. Prediction results of Branch 1.
Figure 8. Prediction results of Branch 1.
Sensors 23 03570 g008
Figure 9. Prediction results of Branch 2.
Figure 9. Prediction results of Branch 2.
Sensors 23 03570 g009
Figure 10. HMM clustering in round 2. Here, red and green colours are used for different clusters.
Figure 10. HMM clustering in round 2. Here, red and green colours are used for different clusters.
Sensors 23 03570 g010
Figure 11. FL results of second round of clustering after 30 communication rounds. (a) FL results when Branch 2 + 4. (b) FL result of Branch 3.
Figure 11. FL results of second round of clustering after 30 communication rounds. (a) FL results when Branch 2 + 4. (b) FL result of Branch 3.
Sensors 23 03570 g011
Figure 12. Prediction result when Branches 2 and 4 were merged together.
Figure 12. Prediction result when Branches 2 and 4 were merged together.
Sensors 23 03570 g012
Figure 13. Prediction result of Branch 3.
Figure 13. Prediction result of Branch 3.
Sensors 23 03570 g013
Figure 14. Graphical elaboration of FedBranched on considered example.
Figure 14. Graphical elaboration of FedBranched on considered example.
Sensors 23 03570 g014
Table 1. MAPE comparison of traditional FL and FedBranched.
Table 1. MAPE comparison of traditional FL and FedBranched.
ClientVanilla FL (%)FedBranched (%)Percentage Improvement (%)
14.132.781.35
22.802.660.14
34.052.981.07
43.633.95−0.32
53.883.650.23
612.432.619.82
715.714.3511.36
82.512.50−0.01
92.852.820.03
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Manzoor, H.U.; Khan, A.R.; Flynn, D.; Alam, M.M.; Akram, M.; Imran, M.A.; Zoha, A. FedBranched: Leveraging Federated Learning for Anomaly-Aware Load Forecasting in Energy Networks. Sensors 2023, 23, 3570. https://doi.org/10.3390/s23073570

AMA Style

Manzoor HU, Khan AR, Flynn D, Alam MM, Akram M, Imran MA, Zoha A. FedBranched: Leveraging Federated Learning for Anomaly-Aware Load Forecasting in Energy Networks. Sensors. 2023; 23(7):3570. https://doi.org/10.3390/s23073570

Chicago/Turabian Style

Manzoor, Habib Ullah, Ahsan Raza Khan, David Flynn, Muhammad Mahtab Alam, Muhammad Akram, Muhammad Ali Imran, and Ahmed Zoha. 2023. "FedBranched: Leveraging Federated Learning for Anomaly-Aware Load Forecasting in Energy Networks" Sensors 23, no. 7: 3570. https://doi.org/10.3390/s23073570

APA Style

Manzoor, H. U., Khan, A. R., Flynn, D., Alam, M. M., Akram, M., Imran, M. A., & Zoha, A. (2023). FedBranched: Leveraging Federated Learning for Anomaly-Aware Load Forecasting in Energy Networks. Sensors, 23(7), 3570. https://doi.org/10.3390/s23073570

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop