TY - JOUR
T1 - SpaceDML
T2 - Enabling Distributed Machine Learning in Space Information Networks
AU - Guo, Hanxi
AU - Yang, Qing
AU - Wang, Hao
AU - Hua, Yang
AU - Song, Tao
AU - Ma, Ruhui
AU - Guan, Haibing
N1 - Publisher Copyright:
© 1986-2012 IEEE.
PY - 2021/7/1
Y1 - 2021/7/1
N2 - Space information networks (SINs) have become a rapidly growing global infrastructure service. Massive volumes of high-resolution images and videos captured by low-orbit satellites and unmanned aerial vehicles have provided a rich training data source for machine learning applications. However, SIN devices' limited communication and computation resources make it challenging to perform machine learning efficiently with a swarm of SIN devices. In this article, we propose Spacedml, a distributed machine learning system for SIN platforms that applies dynamic model compression techniques to adapt distributed machine learning training to SINs' limited bandwidth and unstable connectivity. Spaced-ml has two key algorithm:s adaptive loss-aware quantization, which compresses models without sacrificing their quality, and partial weight averaging, which selectively averages active clients' partial model updates. These algorithms jointly improve communication efficiency and enhance the scalability of distributed machine learning with SIN devices. We evaluate Spacedml by training a LeNet-S model on the MNIST dataset. The experimental results show that Spacedml can increase model accuracy by 2-3 percent and reduce communication bandwidth consumption by up to 60 percent compared to the baseline algorithm.
AB - Space information networks (SINs) have become a rapidly growing global infrastructure service. Massive volumes of high-resolution images and videos captured by low-orbit satellites and unmanned aerial vehicles have provided a rich training data source for machine learning applications. However, SIN devices' limited communication and computation resources make it challenging to perform machine learning efficiently with a swarm of SIN devices. In this article, we propose Spacedml, a distributed machine learning system for SIN platforms that applies dynamic model compression techniques to adapt distributed machine learning training to SINs' limited bandwidth and unstable connectivity. Spaced-ml has two key algorithm:s adaptive loss-aware quantization, which compresses models without sacrificing their quality, and partial weight averaging, which selectively averages active clients' partial model updates. These algorithms jointly improve communication efficiency and enhance the scalability of distributed machine learning with SIN devices. We evaluate Spacedml by training a LeNet-S model on the MNIST dataset. The experimental results show that Spacedml can increase model accuracy by 2-3 percent and reduce communication bandwidth consumption by up to 60 percent compared to the baseline algorithm.
UR - http://www.scopus.com/inward/record.url?scp=85113411312&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85113411312&partnerID=8YFLogxK
U2 - 10.1109/MNET.011.2100075
DO - 10.1109/MNET.011.2100075
M3 - Review article
AN - SCOPUS:85113411312
SN - 0890-8044
VL - 35
SP - 82
EP - 87
JO - IEEE Network
JF - IEEE Network
IS - 4
M1 - 9520325
ER -