Technique Smooths Path for ‘Federated Learning’ AI Training in Wireless Devices
For Immediate Release
Federated learning is a great tool for training artificial intelligence (AI) systems while protecting data privacy, but the amount of data traffic involved has made it unwieldy for systems that include wireless devices. A new technique uses compression to drastically reduce the size of data transmissions, creating additional opportunities for AI training on wireless technologies.
Federated learning is a form of machine learning involving multiple devices, called clients. Each of the clients is trained using different data and develops its own model for performing a specific task. The clients then send their models to a centralized server. The centralized server draws on each of those models to create a hybrid model, which performs better than any of the other models on their own. The central server then sends this hybrid model back to each of the clients. The entire process is then repeated, with each iteration leading to model updates that ultimately improve the system’s performance.
“One of the advantages of federated learning is that it can allow the overall AI system to improve its performance without compromising the privacy of the data being used to train the system,” says Chau-Wai Wong, co-author of a paper on the new technique and an assistant professor of electrical and computer engineering at North Carolina State University. “For example, you could draw on privileged patient data from multiple hospitals in order to improve diagnostic AI tools, without the hospitals having access to data on each other’s patients.”
There are many tasks that could be improved by drawing on data stored on people’s personal devices, such as smartphones. And federated learning would be a way to make use of that data without compromising anyone’s privacy. However, there’s a stumbling block: federated learning requires a lot of communication between the clients and the central server during training, as they send model updates back and forth. In areas where there is limited bandwidth, or where there is a significant amount of data traffic, the communication between clients and the centralized server can clog wireless connections, making the process slow.
“We were trying to think of a way to expedite wireless communication for federated learning, and drew inspiration from the decades of work that has been done on video compression to develop an improved way of compressing data,” Wong says.
Specifically, the researchers developed a technique that allows the clients to compress data into much smaller packets. The packets are condensed before being sent, and then reconstructed by the centralized server. The process is made possible by a series of algorithms developed by the research team. Using the technique, the researchers were able to condense the amount of wireless data shipped from the clients by as much as 99%. Data sent from the server to the clients is not compressed.
“Our technique makes federated learning viable for wireless devices where there is limited available bandwidth,” says Kai Yue, lead author of the paper and a Ph.D. student at NC State. “For example, it could be used to improve the performance of many AI programs that interface with users, such as voice-activated virtual assistants.”
The paper, “Communication-Efficient Federated Learning via Predictive Coding,” is published in the IEEE Journal of Selected Topics in Signal Processing. The paper was co-authored by Huaiyu Dai, a professor of electrical and computer engineering at NC State; and by Richeng Jin, a postdoctoral researcher at NC State. The work was done with partial support from the National Science Foundation, under grant number 1824518.
Note to Editors: The study abstract follows.
“Communication-Efficient Federated Learning via Predictive Coding”
Authors: Kai Yue, Richeng Jin, Chau-Wai Wong and Huaiyu Dai, North Carolina State University
Published: Jan. 13, IEEE Journal of Selected Topics in Signal Processing
Abstract: Federated learning can enable remote workers to collaboratively train a shared machine learning model while allowing training data to be kept locally. In the use case of wireless mobile devices, the communication overhead is a critical bottleneck due to limited power and bandwidth. Prior work has utilized various data compression tools such as quantization and sparsification to reduce the overhead. In this paper, we propose a predictive coding based communication scheme for federated learning. The scheme has shared prediction functions among all devices and allows each worker to transmit a compressed residual vector derived from the reference. In each communication round, we select the predictor and quantizer based on the rate-distortion cost, and further reduce redundancy by using the entropy coding. Extensive simulations reveal that the communication cost can be reduced up to 99% with better learning performance when compared with other baselines.