Concepedia

Publication | Closed Access

Federated Learning over Wireless Networks: Optimization Model Design and Analysis

995

Citations

14

References

2019

Year

TLDR

Federated Learning distributes model training across mobile user equipments, offering data privacy and leveraging many powerful devices, yet challenges such as wireless channel uncertainty, heterogeneous power limits, and varying local data sizes create trade‑offs between computation‑communication latency, learning time, and UE energy consumption. This work formulates the Federated Learning over wireless networks as an optimization problem, FEDL, that explicitly captures these trade‑offs. Although FEDL is non‑convex, the authors decompose it into three convex sub‑problems, enabling tractable analysis. Closed‑form solutions to the sub‑problems yield the globally optimal FEDL learning time, accuracy, and UE energy cost, and numerical experiments confirm the theoretical insights.

Abstract

There is an increasing interest in a new machine learning technique called Federated Learning, in which the model training is distributed over mobile user equipments (UEs), and each UE contributes to the learning model by independently computing the gradient based on its local training data. Federated Learning has several benefits of data privacy and potentially a large amount of UE participants with modern powerful processors and low-delay mobile-edge networks. While most of the existing work focused on designing learning algorithms with provable convergence time, other issues such as uncertainty of wireless channels and UEs with heterogeneous power constraints and local data size, are under-explored. These issues especially affect to various trade-offs: (i) between computation and communication latencies determined by learning accuracy level, and thus (ii) between the Federated Learning time and UE energy consumption. We fill this gap by formulating a Federated Learning over wireless network as an optimization problem FEDL that captures both trade-offs. Even though FEDL is non-convex, we exploit the problem structure to decompose and transform it to three convex sub-problems. We also obtain the globally optimal solution by charactering the closed-form solutions to all sub-problems, which give qualitative insights to problem design via the obtained optimal FEDL learning time, accuracy level, and UE energy cost. Our theoretical analysis is also illustrated by extensive numerical results.

References

YearCitations

Page 1