Echo-CGC: A Communication-Efficient Byzantine-Tolerant Distributed Machine Learning Algorithm in Single-Hop Radio Network
In the past few years, many Byzantine-tolerant distributed machine learning (DML) algorithms have been proposed in the point-to-point communication model. In this paper, we focus on a popular DML framework - the parameter server computation paradigm and iterative learning algorithms that proceed in...
Main Authors: | , |
---|---|
Format: | Conference Object |
Language: | English |
Published: |
LIPIcs - Leibniz International Proceedings in Informatics. 24th International Conference on Principles of Distributed Systems (OPODIS 2020)
2021
|
Subjects: | |
Online Access: | https://doi.org/10.4230/LIPIcs.OPODIS.2020.7 https://nbn-resolving.org/urn:nbn:de:0030-drops-134927 https://drops.dagstuhl.de/opus/volltexte/2021/13492/ |
id |
ftdagstuhl:oai:drops-oai.dagstuhl.de:13492 |
---|---|
record_format |
openpolar |
spelling |
ftdagstuhl:oai:drops-oai.dagstuhl.de:13492 2023-05-15T16:01:33+02:00 Echo-CGC: A Communication-Efficient Byzantine-Tolerant Distributed Machine Learning Algorithm in Single-Hop Radio Network Zhang, Qinzi Tseng, Lewis 2021 application/pdf https://doi.org/10.4230/LIPIcs.OPODIS.2020.7 https://nbn-resolving.org/urn:nbn:de:0030-drops-134927 https://drops.dagstuhl.de/opus/volltexte/2021/13492/ eng eng LIPIcs - Leibniz International Proceedings in Informatics. 24th International Conference on Principles of Distributed Systems (OPODIS 2020) doi:10.4230/LIPIcs.OPODIS.2020.7 urn:nbn:de:0030-drops-134927 https://drops.dagstuhl.de/opus/volltexte/2021/13492/ https://creativecommons.org/licenses/by/3.0/ CC-BY Distributed Machine Learning Single-hop Radio Network Byzantine Fault Communication Complexity Wireless Communication Parameter Server Data processing Computer science InProceedings publishedVersion 2021 ftdagstuhl https://doi.org/10.4230/LIPIcs.OPODIS.2020.7 2022-05-23T15:20:38Z In the past few years, many Byzantine-tolerant distributed machine learning (DML) algorithms have been proposed in the point-to-point communication model. In this paper, we focus on a popular DML framework - the parameter server computation paradigm and iterative learning algorithms that proceed in rounds, e.g., [Gupta and Vaidya, 2020; El-Mhamdi et al., 2020; Chen et al., 2017]. One limitation of prior algorithms in this domain is the high communication complexity. All the Byzantine-tolerant DML algorithms that we are aware of need to send n d-dimensional vectors from worker nodes to the parameter server in each round, where n is the number of workers and d is the number of dimensions of the feature space (which may be in the order of millions). In a wireless network, power consumption is proportional to the number of bits transmitted. Consequently, it is extremely difficult, if not impossible, to deploy these algorithms in power-limited wireless devices. Motivated by this observation, we aim to reduce the communication complexity of Byzantine-tolerant DML algorithms in the single-hop radio network [Alistarh et al., 2010; Bhandari and Vaidya, 2005; Koo, 2004]. Inspired by the CGC filter developed by Gupta and Vaidya, PODC 2020 [Gupta and Vaidya, 2020], we propose a gradient descent-based algorithm, Echo-CGC. Our main novelty is a mechanism to utilize the broadcast properties of the radio network to avoid transmitting the raw gradients (full d-dimensional vectors). In the radio network, each worker is able to overhear previous gradients that were transmitted to the parameter server. Roughly speaking, in Echo-CGC, if a worker "agrees" with a combination of prior gradients, it will broadcast the "echo message" instead of the its raw local gradient. The echo message contains a vector of coefficients (of size at most n) and the ratio of the magnitude between two gradients (a float). In comparison, the traditional approaches need to send n local gradients in each round, where each gradient is typically a vector in a ... Conference Object DML DROPS - Dagstuhl Research Online Publication Server (Schloss Dagstuhl - Leibniz Center for Informatics ) |
institution |
Open Polar |
collection |
DROPS - Dagstuhl Research Online Publication Server (Schloss Dagstuhl - Leibniz Center for Informatics ) |
op_collection_id |
ftdagstuhl |
language |
English |
topic |
Distributed Machine Learning Single-hop Radio Network Byzantine Fault Communication Complexity Wireless Communication Parameter Server Data processing Computer science |
spellingShingle |
Distributed Machine Learning Single-hop Radio Network Byzantine Fault Communication Complexity Wireless Communication Parameter Server Data processing Computer science Zhang, Qinzi Tseng, Lewis Echo-CGC: A Communication-Efficient Byzantine-Tolerant Distributed Machine Learning Algorithm in Single-Hop Radio Network |
topic_facet |
Distributed Machine Learning Single-hop Radio Network Byzantine Fault Communication Complexity Wireless Communication Parameter Server Data processing Computer science |
description |
In the past few years, many Byzantine-tolerant distributed machine learning (DML) algorithms have been proposed in the point-to-point communication model. In this paper, we focus on a popular DML framework - the parameter server computation paradigm and iterative learning algorithms that proceed in rounds, e.g., [Gupta and Vaidya, 2020; El-Mhamdi et al., 2020; Chen et al., 2017]. One limitation of prior algorithms in this domain is the high communication complexity. All the Byzantine-tolerant DML algorithms that we are aware of need to send n d-dimensional vectors from worker nodes to the parameter server in each round, where n is the number of workers and d is the number of dimensions of the feature space (which may be in the order of millions). In a wireless network, power consumption is proportional to the number of bits transmitted. Consequently, it is extremely difficult, if not impossible, to deploy these algorithms in power-limited wireless devices. Motivated by this observation, we aim to reduce the communication complexity of Byzantine-tolerant DML algorithms in the single-hop radio network [Alistarh et al., 2010; Bhandari and Vaidya, 2005; Koo, 2004]. Inspired by the CGC filter developed by Gupta and Vaidya, PODC 2020 [Gupta and Vaidya, 2020], we propose a gradient descent-based algorithm, Echo-CGC. Our main novelty is a mechanism to utilize the broadcast properties of the radio network to avoid transmitting the raw gradients (full d-dimensional vectors). In the radio network, each worker is able to overhear previous gradients that were transmitted to the parameter server. Roughly speaking, in Echo-CGC, if a worker "agrees" with a combination of prior gradients, it will broadcast the "echo message" instead of the its raw local gradient. The echo message contains a vector of coefficients (of size at most n) and the ratio of the magnitude between two gradients (a float). In comparison, the traditional approaches need to send n local gradients in each round, where each gradient is typically a vector in a ... |
format |
Conference Object |
author |
Zhang, Qinzi Tseng, Lewis |
author_facet |
Zhang, Qinzi Tseng, Lewis |
author_sort |
Zhang, Qinzi |
title |
Echo-CGC: A Communication-Efficient Byzantine-Tolerant Distributed Machine Learning Algorithm in Single-Hop Radio Network |
title_short |
Echo-CGC: A Communication-Efficient Byzantine-Tolerant Distributed Machine Learning Algorithm in Single-Hop Radio Network |
title_full |
Echo-CGC: A Communication-Efficient Byzantine-Tolerant Distributed Machine Learning Algorithm in Single-Hop Radio Network |
title_fullStr |
Echo-CGC: A Communication-Efficient Byzantine-Tolerant Distributed Machine Learning Algorithm in Single-Hop Radio Network |
title_full_unstemmed |
Echo-CGC: A Communication-Efficient Byzantine-Tolerant Distributed Machine Learning Algorithm in Single-Hop Radio Network |
title_sort |
echo-cgc: a communication-efficient byzantine-tolerant distributed machine learning algorithm in single-hop radio network |
publisher |
LIPIcs - Leibniz International Proceedings in Informatics. 24th International Conference on Principles of Distributed Systems (OPODIS 2020) |
publishDate |
2021 |
url |
https://doi.org/10.4230/LIPIcs.OPODIS.2020.7 https://nbn-resolving.org/urn:nbn:de:0030-drops-134927 https://drops.dagstuhl.de/opus/volltexte/2021/13492/ |
genre |
DML |
genre_facet |
DML |
op_relation |
doi:10.4230/LIPIcs.OPODIS.2020.7 urn:nbn:de:0030-drops-134927 https://drops.dagstuhl.de/opus/volltexte/2021/13492/ |
op_rights |
https://creativecommons.org/licenses/by/3.0/ |
op_rightsnorm |
CC-BY |
op_doi |
https://doi.org/10.4230/LIPIcs.OPODIS.2020.7 |
_version_ |
1766397362944606208 |