Abstract

Scalable and sustainable AI-driven analytics are necessary to enable large-scale and heterogeneous service deployment in sixth-generation (6G) ultra-dense networks. This implies that the exchange of raw monitoring data should be minimized across the network by bringing the analysis functions closer to the data collection points. While federated learning (FL) is an efficient tool to implement such a decentralized strategy, real networks are generally characterized by time- and space-varying traffic patterns and channel conditions, making thereby the data collected in different points non independent and identically distributed (non-IID), which is challenging for FL. To sidestep this issue, we first introduce a new a priori metric that we call dataset entropy, whose role is to capture the distribution, the quantity of information, the unbalanced structure and the “non-IIDness” of a dataset independently of the models. This a priori entropy is calculated using a multi-dimensional spectral clustering scheme over both the features and the supervised output spaces, and is suitable for classification as well as regression tasks. The FL aggregation operations support system (OSS) server then uses the reported dataset entropies to devise 1) an entropy-based federated averaging scheme, and 2) a stochastic participant selection policy to significantly stabilize the training, minimize the convergence time, and reduce the corresponding computation cost. Numerical results are provided to show the superiority of these novel approaches.

Highlights

  • 1.1 Related WorkIn (Brendan McMahan et al, 2017), the authors have proposed the federated averaging (FedAvg) algorithm that synchronously aggregates the parameters, and is susceptible to the so-calledEntropy-Driven Stochastic Federated Learning straggler effect, i.e., each training round only progresses as fast as the slowest edge device since the FL server waits for all devices to complete local training before the global aggregation can be performed

  • Since the datasets of central unit (CU) with low entropy can hold samples that are non-existing in the other high entropy datasets and yet can help in further generalizing the FL model, the idea we have proposed is to give them a chance by implementing a softmax stochastic policy, where each CU can participate in the training with a probability proportional to its entropy

  • Knowing how critical is the bandwidth occupation for FL exchanges, and how the CUs local model training is power consuming, especially in 6G mobile systems, our introduced entropy stochastic policy shows good results

Read more

Summary

Related Work

In (Brendan McMahan et al, 2017), the authors have proposed the federated averaging (FedAvg) algorithm that synchronously aggregates the parameters, and is susceptible to the so-called. The model convergence is found to be significantly delayed when data is non independent and identically distributed (non-IID) and unbalanced (Zhao et al, 2018) To solve this issue, it has been proposed to distribute a public dataset to the FL clients at the beginning. The algorithm is still unable to generalize to suit the dynamic computation constraints of heterogeneous devices Given this uncertainty surrounding the reliability of asynchronous FL, synchronous FL remains the most commonly used approach (Keith et al, 2019). In this context, it has been confirmed that the correlation between the model parameters of different clients is increasing as the training progresses, which implies that aggregating parameters directly by averaging may not be a reasonable approach in general (Xiao et al, 2020). As energy consumption is one of the important aspects to consider in FL, in (Tran et al, 2019) the trade-off between learning time, learning accuracy and terminals power consumption has been investigated

Contributions
Edge-RAN
Data Collection
PROPOSED ENTROPY-BASED FEDERATED LEARNING
Dataset Entropy
Entropy-Driven FL Combining
Entropy-Driven Stochastic FL Policy
STOCHASTIC FEDERATED LEARNING CONVERGENCE ANALYSIS
Settings and Baselines
Numerical Results Analysis
CONCLUSION
Full Text
Published version (Free)

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call