Abstract

Federated Learning (FL) is a promising technique to enhance the safety and efficiency of intelligent transportation systems. While FL has been extensively studied, the communication and networking challenges related to the operations of FL in dynamic yet dense vehicular networks remain under-explored. Limited storage and communication capacities of individual vehicles throttle the timely training of an FL model in distributed vehicular networks. In this paper, we present a communication framework for FL (CF4FL) in transportation systems. CF4FL aims to accelerate the convergence of FL training process through the innovation of two complementary networking components: (i) a deadline-driven vehicle scheduler (DDVS), and (ii) a concurrent vehicle polling scheme (CVPS). DDVS identifies a subset of vehicles for local model training in each iteration of FL, with the aim of minimizing data loss while respecting the deadline constraints derived from vehicles’ storage, computation, and energy budgets. CVPS takes advantage of multiple antennas on an edge server to enable concurrent local model transmissions in dynamic vehicular networks, thereby reducing the airtime overhead of each FL iteration. We have evaluated CF4FL through a blend of experimentation and simulation. Trace-driven simulation shows that, compared to existing scheduling and transmission schemes, CF4FL reduces the convergence time of FL training by 39%.

Full Text
Published version (Free)

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call