FedQMIX: Communication-efficient federated learning via multi-agent reinforcement learning

  • Abstract
  • Literature Map
  • Similar Papers
Abstract
Translate article icon Translate Article Star icon
Take notes icon Take Notes

FedQMIX: Communication-efficient federated learning via multi-agent reinforcement learning

Similar Papers
  • Research Article
  • Cite Count Icon 139
  • 10.1109/tccn.2021.3084406
Fast-Convergent Federated Learning With Adaptive Weighting
  • Dec 1, 2021
  • IEEE Transactions on Cognitive Communications and Networking
  • Hongda Wu + 1 more

Federated learning (FL) enables resource-constrained edge nodes to collaboratively learn a global model under the orchestration of a central server while keeping privacy-sensitive data locally. The non-independent-and-identically-distributed (non-IID) data samples across participating nodes slow model training and impose additional communication rounds for FL to converge. In this paper, we propose <monospace xmlns:mml="http://www.w3.org/1998/Math/MathML" xmlns:xlink="http://www.w3.org/1999/xlink">Fed</monospace> erated <monospace xmlns:mml="http://www.w3.org/1998/Math/MathML" xmlns:xlink="http://www.w3.org/1999/xlink">Ad</monospace> a <monospace xmlns:mml="http://www.w3.org/1998/Math/MathML" xmlns:xlink="http://www.w3.org/1999/xlink">p</monospace> tive Weighting ( <monospace xmlns:mml="http://www.w3.org/1998/Math/MathML" xmlns:xlink="http://www.w3.org/1999/xlink">FedAdp</monospace> ) algorithm that aims to accelerate model convergence under the presence of nodes with non-IID dataset. We observe the implicit connection between the node contribution to the global model aggregation and data distribution on the local node through theoretical and empirical analysis. We then propose to assign different weights for updating the global model based on node contribution adaptively through each training round. The contribution of participating nodes is first measured by the angle between the local gradient vector and the global gradient vector, and then, weight is quantified by a designed non-linear mapping function subsequently. The simple yet effective strategy can reinforce positive (suppress negative) node contribution dynamically, resulting in communication round reduction drastically. Its superiority over the commonly adopted Federated Averaging ( <monospace xmlns:mml="http://www.w3.org/1998/Math/MathML" xmlns:xlink="http://www.w3.org/1999/xlink">FedAvg</monospace> ) is verified both theoretically and experimentally. With extensive experiments performed in Pytorch and PySyft, we show that FL training with <monospace xmlns:mml="http://www.w3.org/1998/Math/MathML" xmlns:xlink="http://www.w3.org/1999/xlink">FedAdp</monospace> can reduce the number of communication rounds by up to 54.1% on MNIST dataset and up to 45.4% on FashionMNIST dataset, as compared to <monospace xmlns:mml="http://www.w3.org/1998/Math/MathML" xmlns:xlink="http://www.w3.org/1999/xlink">FedAvg</monospace> algorithm.

  • Conference Article
  • Cite Count Icon 9
  • 10.1109/icc42927.2021.9500890
Fast-Convergent Federated Learning with Adaptive Weighting
  • Jun 1, 2021
  • Hongda Wu + 1 more

Federated learning (FL) enables resource-constrained edge nodes to collaboratively learn a global model under the orchestration of a central server while keeping privacy-sensitive data locally. The non-independent-and-identically-distributed (non-IID) data samples across participating nodes slow model training and impose additional communication rounds for FL to converge. In this paper, we propose Fed erated Adaptive Weighting (FedAdp) algorithm that aims to accelerate model convergence under the presence of nodes with non-IID dataset. Through mathematical and empirical analysis, we observe the implicit connection between the gradient of local training and data distribution on local node. We then propose to assign different weight for updating global model based on node contribution adaptively through each training round, which is measured by the angle between local gradient vector and global gradient vector, and is quantified by a designed non-linear mapping function. The simple yet effective strategy can reinforce positive (suppress negative) node contribution dynamically, that results in communication round reduction drastically. With extensive experiments performed in Pytorch and PySyft, we show that FL training with FedAdp can reduce the number of communication rounds by up to 54.1% on MNIST dataset and up to 45.4% on FashionMNIST dataset, as compared to the commonly adopted Federated Averaging (FedAvg) algorithm.

  • Research Article
  • Cite Count Icon 20
  • 10.1109/tcyb.2023.3247365
FedLGA: Toward System-Heterogeneity of Federated Learning via Local Gradient Approximation.
  • Jan 1, 2024
  • IEEE Transactions on Cybernetics
  • Xingyu Li + 3 more

Federated learning (FL) is a decentralized machine learning architecture, which leverages a large number of remote devices to learn a joint model with distributed training data. However, the system-heterogeneity is one major challenge in an FL network to achieve robust distributed learning performance, which comes from two aspects: 1) device-heterogeneity due to the diverse computational capacity among devices and 2) data-heterogeneity due to the nonidentically distributed data across the network. Prior studies addressing the heterogeneous FL issue, for example, FedProx, lack formalization and it remains an open problem. This work first formalizes the system-heterogeneous FL problem and proposes a new algorithm, called federated local gradient approximation (FedLGA), to address this problem by bridging the divergence of local model updates via gradient approximation. To achieve this, FedLGA provides an alternated Hessian estimation method, which only requires extra linear complexity on the aggregator. Theoretically, we show that with a device-heterogeneous ratio ρ , FedLGA achieves convergence rates on non-i.i.d. distributed FL training data for the nonconvex optimization problems with O ([(1+ρ)/√{ENT}] + 1/T) and O ([(1+ρ)√E/√{TK}] + 1/T) for full and partial device participation, respectively, where E is the number of local learning epoch, T is the number of total communication round, N is the total device number, and K is the number of the selected device in one communication round under partially participation scheme. The results of comprehensive experiments on multiple datasets indicate that FedLGA can effectively address the system-heterogeneous problem and outperform current FL methods. Specifically, the performance against the CIFAR-10 dataset shows that, compared with FedAvg, FedLGA improves the model's best testing accuracy from 60.91% to 64.44%.

  • Research Article
  • Cite Count Icon 29
  • 10.1109/jiot.2022.3183295
Joint Resource Allocation to Minimize Execution Time of Federated Learning in Cell-Free Massive MIMO
  • Nov 1, 2022
  • IEEE Internet of Things Journal
  • Tung Thanh Vu + 5 more

Due to its communication efficiency and privacy-preserving capability, federated learning (FL) has emerged as a promising framework for machine learning in 5G-and-beyond wireless networks. Of great interest is the design and optimization of new wireless network structures that support the stable and fast operation of FL. Cell-free massive multiple-input–multiple-output (CFmMIMO) turns out to be a suitable candidate, which allows each communication round in the iterative FL process to be stably executed within a large-scale coherence time. Aiming to reduce the total execution time of the FL process in CFmMIMO, this article proposes choosing only a subset of available users to participate in FL. An optimal selection of users with favorable link conditions would minimize the execution time of each communication round while limiting the total number of communication rounds required. Toward this end, we formulate a joint optimization problem of user selection, transmit power, and processing frequency, subject to a predefined minimum number of participating users to guarantee the quality of learning. We then develop a new algorithm that is proven to converge to the neighborhood of the stationary points of the formulated problem. Numerical results confirm that our proposed approach significantly reduces the FL total execution time over baseline schemes. The time reduction is more pronounced when the density of access point deployments is moderately low.

  • Research Article
  • Cite Count Icon 35
  • 10.1609/aaai.v37i8.26177
Securing Secure Aggregation: Mitigating Multi-Round Privacy Leakage in Federated Learning
  • Jun 26, 2023
  • Proceedings of the AAAI Conference on Artificial Intelligence
  • Jinhyun So + 4 more

Secure aggregation is a critical component in federated learning (FL), which enables the server to learn the aggregate model of the users without observing their local models. Conventionally, secure aggregation algorithms focus only on ensuring the privacy of individual users in a single training round. We contend that such designs can lead to significant privacy leakages over multiple training rounds, due to partial user selection/participation at each round of FL. In fact, we show that the conventional random user selection strategies in FL lead to leaking users' individual models within number of rounds that is linear in the number of users. To address this challenge, we introduce a secure aggregation framework, Multi-RoundSecAgg, with multi-round privacy guarantees. In particular, we introduce a new metric to quantify the privacy guarantees of FL over multiple training rounds, and develop a structured user selection strategy that guarantees the long-term privacy of each user (over any number of training rounds). Our framework also carefully accounts for the fairness and the average number of participating users at each round. Our experiments on MNIST, CIFAR-10 and CIFAR-100 datasets in the IID and the non-IID settings demonstrate the performance improvement over the baselines, both in terms of privacy protection and test accuracy.

  • Research Article
  • 10.13052/jwe1540-9589.2381
Personalized User Models in a Real-world Edge Computing Environment: A Peer-to-peer Federated Learning Framework
  • Feb 7, 2025
  • Journal of Web Engineering
  • Xiangchi Song + 3 more

As the number of IoT devices and the volume of data increase, distributed computing systems have become the primary deployment solution for large-scale Internet of Things (IoT) environments. Federated learning (FL) is a collaborative machine learning framework that allows for model training using data from all participants while protecting their privacy. However, traditional FL suffers from low computational and communication efficiency in large-scale hierarchical cloud-edge collaborative IoT systems. Additionally, due to heterogeneity issues, not all IoT devices necessarily benefit from the global model of traditional FL, but instead require the maintenance of personalized levels in the global training process. Therefore we extend FL into a horizontal peer-to-peer (P2P) structure and introduce our P2PFL framework: efficient peer-to-peer federated learning for users (EPFLU). EPFLU transitions the paradigms from vertical FL to a horizontal P2P structure from the user perspective and incorporates personalized enhancement techniques using private information. Through horizontal consensus information aggregation and private information supplementation, EPFLU solves the weakness of traditional FL that dilutes the characteristics of individual client data and leads to model deviation. This structural transformation also significantly alleviates the original communication issues. Additionally, EPFLU has a customized simulation evaluation framework, and uses the EUA dataset containing real-world edge server distribution, making it more suitable for real-world large-scale IoT. Within this framework, we design two extreme data distribution scenarios and conduct detailed experiments of EPFLU and selected baselines on the MNIST and CIFAR-10 datasets. The results demonstrate that the robust and adaptive EPFLU framework can consistently converge to optimal performance even under challenging data distribution scenarios. Compared with the traditional FL and selected P2PFL methods, EPFLU achieves communication time improvements of 39% and 16% respectively.

  • Research Article
  • 10.24908/iqurcp18058
Enhancing Transmission Efficiency in Non-IID Federated Learning by Integrating Quantization into an Existing Trustworthiness Model
  • Sep 9, 2024
  • Inquiry@Queen's Undergraduate Research Conference Proceedings
  • Zain Parihar

Federated Learning (FL) is a distributed machine learning paradigm that enables multiple clients to collaboratively train a model without sharing their data, thus preserving privacy. In this approach, data is not transmitted between the client and the base station. Instead, a model is sent to each device, where it is trained locally, and then retransmitted back to the base station. Each iteration of this process is known as a communication round. However, FL faces challenges, especially when data is not independently and identically distributed (Non-IID). Non-IID data means that the data across clients can vary significantly in distribution, leading to situations where certain classes or features are overrepresented in some clients and underrepresented in others. This lack of uniformity quickly leads to biased model updates and reduced performance. To address this, previous studies have introduced trustworthiness metrics to ensure more reliable model aggregation, minimizing accuracy losses associated with Non-IID data. Another significant challenge is the high transmission load, as model updates between clients and the base station are resource-intensive. Model parameters are transmitted between clients and base stations, which can strain communication channels and slow down the entire process, especially when dealing with larger models and datasets. This communication overhead is a bottleneck that limits the scalability of FL, particularly in resource-constrained environments. Our research addresses these challenges by integrating quantization into the trustworthiness model specifically in a Non-IID scenario. Quantization reduces the precision of model parameters to minimize data transmission, resulting in more efficient communication. We applied different levels of quantization intensity to a model trained on the CIFAR-10 dataset and found that certain methods can significantly reduce transmission overhead without substantial accuracy loss. Our findings suggest that combining quantization with trustworthiness metrics can significantly enhance the efficiency and potentially improve the adoption of Federated Learning in resource-constrained environments.

  • Research Article
  • 10.34028/iajit/22/6/1
Strategic Optimization of Convergence and Energy in Federated Learning Systems
  • Jan 1, 2025
  • The International Arab Journal of Information Technology
  • Ghassan Samara + 8 more

Federated Learning (FL) is a Machine Learning (ML) paradigm in which multiple devices collaboratively train a model without sharing their local data. This decentralized approach provides significant privacy benefits, enabling compliance with data protection regulations and safeguarding sensitive user information by keeping raw data on local devices. Instead of transmitting raw data, FL sends model updates to a central aggregator to improve the global model. However, this process can result in higher Carbon Dioxide (CO₂) emissions compared to traditional centralized ML systems, due to the increased number of participating devices and communication rounds. This study evaluates the performance, convergence speed, energy efficiency, and environmental impact of FL models compared to centralized models, using the Modified National Institute of Standards and Technology dataset (MNIST) and Canadian Institute for Advanced Research-10 classes dataset (CIFAR-10). Four models were tested: two FL models and two centralized models. The evaluation focused on accuracy, number of training rounds to convergence, and total CO₂ emissions. To optimize both convergence and energy efficiency, a dynamic hill-climbing-based early stopping technique was introduced. After every 100 rounds, model accuracy improvements were assessed, and training was terminated early if further gains fell below a shrinking threshold, effectively reducing unnecessary computation and energy consumption. Results show that, under the tested conditions, FL models achieved competitive or higher accuracy than centralized models, particularly on non-Independent and Identically Distributed (IID) data distributions. For example, the federated MNIST model reached 98.79% accuracy with a significantly lower carbon footprint when early stopping was applied. Overall, the proposed optimization approach reduced CO₂ emissions by approximately 60% without substantial loss in accuracy. By integrating privacy preservation, explicit regulatory relevance, and a practical dynamic optimization method, this research demonstrates that FL can deliver strong model performance while meeting modern requirements for data privacy and environmental sustainability

  • Research Article
  • Cite Count Icon 1
  • 10.3390/app14188299
Uncertainty-Aware Federated Reinforcement Learning for Optimizing Accuracy and Energy in Heterogeneous Industrial IoT
  • Sep 14, 2024
  • Applied Sciences
  • A S M Sharifuzzaman Sagar + 3 more

The Internet of Things (IoT) technology has revolutionized various industries by allowing data collection, analysis, and decision-making in real time through interconnected devices. However, challenges arise in implementing Federated Learning (FL) in heterogeneous industrial IoT environments, such as maintaining model accuracy with non-Independent and Identically Distributed (non-IID) datasets and straggler IoT devices, ensuring computation and communication efficiency, and addressing weight aggregation issues. In this study, we propose an Uncertainty-Aware Federated Reinforcement Learning (UA-FedRL) method that dynamically selects epochs of individual clients to effectively manage heterogeneous industrial IoT devices and improve accuracy, computation, and communication efficiency. Additionally, we introduce the Predictive Weighted Average Aggregation (PWA) method to tackle weight aggregation issues in heterogeneous industrial IoT scenarios by adjusting the weights of individual models based on their quality. The UA-FedRL addresses the inherent complexities and challenges of implementing FL in heterogeneous industrial IoT environments. Extensive simulations in complex IoT environments demonstrate the superior performance of UA-FedRL on both MNIST and CIFAR-10 datasets compared to other existing approaches in terms of accuracy, communication efficiency, and computation efficiency. The UA-FedRL algorithm attain an accuracy of 96.83% on the MNIST dataset and 62.75% on the CIFAR-10 dataset, despite the presence of 90% straggler IoT devices, attesting to its robust performance and adaptability in different datasets.

  • Conference Article
  • Cite Count Icon 2
  • 10.1109/smc53654.2022.9945192
FedGosp: A Novel Framework of Gossip Federated Learning for Data Heterogeneity
  • Oct 9, 2022
  • Guanghao Li + 5 more

Federated learning (FL) provides the possibility to solve the problem of data privacy, but it suffers much from the data heterogeneity among different participants. Currently, some promising FL algorithms improve the effectiveness of learning under the non independent-and-identically-distributed (Non-IID) data settings. However, they require a large number of communication rounds between the server and clients for an acceptable accuracy. Inspired by the training paradigm of gossip learning, this paper proposes a new FL framework, named FedGosp. It first classifies the clients into different categories based on the model weights trained by the locally stored data. Then FedGosp utilizes the communication not only between clients and the server, but also between different classes of clients themselves. This training process enables instilling knowledge about various data distributions in the passed models. We evaluate the performance of FedGosp in multiple Non-IID settings on CIFAR10 and MNIST datasets, and compare it with the recently popular algorithms such as SCAFFOLD, FedAvg and FedProx. The experimental results show that FedGosp can improve the model accuracy by 6.53% and save 5.6 × communication costs at best compared to the second-ranked baseline.

  • Conference Article
  • Cite Count Icon 10
  • 10.1109/ccnc49033.2022.9700601
Communication-Loss Trade-Off in Federated Learning: A Distributed Client Selection Algorithm
  • Jan 8, 2022
  • Minoo Hosseinzadeh + 3 more

Mass data generation occurring in the Internet-of-Things (IoT) requires processing to extract meaningful information. Deep learning is commonly used to perform such processing. However, due to the sensitive nature of these data, it is important to consider data privacy. As such, federated learning (FL) has been proposed to address this issue. FL pushes training to the client devices and tasks a central server with aggregating collected model weights to update a global model. However, the transmission of these model weights can be costly, gradually. The trade-off between communicating model weights for aggregation and the loss provided by the global model remains an open problem. In this work, we cast this trade-off problem of client selection in FL as an optimization problem. We then design a Distributed Client Selection (DCS) algorithm that allows client devices to decide to participate in aggregation in hopes of minimizing overall communication cost — while maintaining low loss. We evaluate the performance of our proposed client selection algorithm against standard FL and a state-of-the-art client selection algorithm, called Power-of-Choice (PoC), using CIFAR-10, FMNIST, and MNIST datasets. Our experimental results confirm that our DCS algorithm is able to closely match the loss provided by the standard FL and PoC, while on average reducing the overall communication cost by nearly 32.67% and 44.71% in comparison to standard FL and PoC, respectively.

  • Research Article
  • Cite Count Icon 48
  • 10.1609/aaai.v36i8.20894
A Multi-Agent Reinforcement Learning Approach for Efficient Client Selection in Federated Learning
  • Jun 28, 2022
  • Proceedings of the AAAI Conference on Artificial Intelligence
  • Sai Qian Zhang + 2 more

Federated learning (FL) is a training technique that enables client devices to jointly learn a shared model by aggregating locally computed models without exposing their raw data. While most of the existing work focuses on improving the FL model accuracy, in this paper, we focus on the improving the training efficiency, which is often a hurdle for adopting FL in real world applications. Specifically, we design an efficient FL framework which jointly optimizes model accuracy, processing latency and communication efficiency, all of which are primary design considerations for real implementation of FL. Inspired by the recent success of Multi Agent Reinforcement Learning (MARL) in solving complex control problems, we present FedMarl, a federated learning framework that relies on trained MARL agents to perform efficient run-time client selection. Experiments show that FedMarl can significantly improve model accuracy with much lower processing latency and communication cost.

  • Research Article
  • Cite Count Icon 8
  • 10.1016/j.knosys.2023.110638
EHR privacy preservation using federated learning with DQRE-Scnet for healthcare application domains
  • May 19, 2023
  • Knowledge-Based Systems
  • Om Kumar C.U + 4 more

EHR privacy preservation using federated learning with DQRE-Scnet for healthcare application domains

  • PDF Download Icon
  • Research Article
  • Cite Count Icon 3
  • 10.1007/s10489-024-05521-y
FLCP: federated learning framework with communication-efficient and privacy-preserving
  • May 1, 2024
  • Applied Intelligence
  • Wei Yang + 4 more

Within the federated learning (FL) framework, the client collaboratively trains the model in coordination with a central server, while the training data can be kept locally on the client. Thus, the FL framework mitigates the privacy disclosure and costs related to conventional centralized machine learning. Nevertheless, current surveys indicate that FL still has problems in terms of communication efficiency and privacy risks. In this paper, to solve these problems, we develop an FL framework with communication-efficient and privacy-preserving (FLCP). To realize the FLCP, we design a novel compression algorithm with efficient communication, namely, adaptive weight compression FedAvg (AWC-FedAvg). On the basis of the non-independent and identically distributed (non-IID) and unbalanced data distribution in FL, a specific compression rate is provided for each client, and homomorphic encryption (HE) and differential privacy (DP) are integrated to provide demonstrable privacy protection and maintain the desirability of the model. Therefore, our proposed FLCP smoothly balances communication efficiency and privacy risks, and we prove its security against “honest-but-curious” servers and extreme collusion under the defined threat model. We evaluate the scheme by comparing it with state-of-the-art results on the MNIST and CIFAR-10 datasets. The results show that the FLCP performs better in terms of training efficiency and model accuracy than the baseline method.

  • Research Article
  • Cite Count Icon 16
  • 10.1109/ojcoms.2023.3263962
Federated Learning Resource Optimization and Client Selection for Total Energy Minimization Under Outage, Latency, and Bandwidth Constraints With Partial or No CSI
  • Jan 1, 2023
  • IEEE Open Journal of the Communications Society
  • Mohamed Hany Mahmoud + 3 more

We consider the problem of minimizing the total energy consumption due to the computation and communication tasks of federated learning (FL) under bandwidth and latency constraints. To avoid channel state information (CSI) feedback to the transmitter, we adopt outage probability as an additional constraint in the energy minimization problem. First, we define a feasibility metric based on the system design parameters to exclude slow clients (stragglers). Then, we propose a novel client selection algorithm, after excluding stragglers, based on dividing the remaining clients into clusters, where clients within the same cluster collaborate in a communication round to train their local models. For each communication round, one client cluster is selected in a round-robin fashion. Furthermore, we formulate and solve a resource allocation problem to optimize the transmit power, clock frequency, allocated bandwidth, and communication latency of clients within each cluster to minimize the total energy subject to total bandwidth, latency, and outage constraints. Moreover, we extend our FL design framework to the case of no CSI at both the client and server ends using differential transmission to eliminate CSI estimation pilot overhead and complexity at comparable total energy consumption and learning accuracy to coherent transmission. We test our proposed algorithms over MNIST and Fashion-MNIST datasets in iid and non-iid settings. Our proposed client selection algorithm reduces the number of participating clients per communication round by 41% compared to the baseline while maintaining the same learning accuracy. Moreover, our results demonstrate that increasing the number of receive antennas at the server from one to four can reduce the number of communication rounds required to reach a predetermined testing accuracy level by up to 53% for the Fashion-MNIST dataset.

Save Icon
Up Arrow
Open/Close
  • Ask R Discovery Star icon
  • Chat PDF Star icon

AI summaries and top papers from 250M+ research sources.

Search IconWhat is the difference between bacteria and viruses?
Open In New Tab Icon
Search IconWhat is the function of the immune system?
Open In New Tab Icon
Search IconCan diabetes be passed down from one generation to the next?
Open In New Tab Icon