Echo-CGC: A Communication-Efficient Byzantine-Tolerant Distributed Machine Learning Algorithm in Single-Hop Radio Network

In the past few years, many Byzantine-tolerant distributed machine learning (DML) algorithms have been proposed in the point-to-point communication model. In this paper, we focus on a popular DML framework - the parameter server computation paradigm and iterative learning algorithms that proceed in...

Full description

Bibliographic Details
Main Authors: Zhang, Qinzi, Tseng, Lewis
Format: Conference Object
Language:English
Published: LIPIcs - Leibniz International Proceedings in Informatics. 24th International Conference on Principles of Distributed Systems (OPODIS 2020) 2021
Subjects:
DML
Online Access:https://doi.org/10.4230/LIPIcs.OPODIS.2020.7
https://nbn-resolving.org/urn:nbn:de:0030-drops-134927
https://drops.dagstuhl.de/opus/volltexte/2021/13492/
id ftdagstuhl:oai:drops-oai.dagstuhl.de:13492
record_format openpolar
spelling ftdagstuhl:oai:drops-oai.dagstuhl.de:13492 2023-05-15T16:01:33+02:00 Echo-CGC: A Communication-Efficient Byzantine-Tolerant Distributed Machine Learning Algorithm in Single-Hop Radio Network Zhang, Qinzi Tseng, Lewis 2021 application/pdf https://doi.org/10.4230/LIPIcs.OPODIS.2020.7 https://nbn-resolving.org/urn:nbn:de:0030-drops-134927 https://drops.dagstuhl.de/opus/volltexte/2021/13492/ eng eng LIPIcs - Leibniz International Proceedings in Informatics. 24th International Conference on Principles of Distributed Systems (OPODIS 2020) doi:10.4230/LIPIcs.OPODIS.2020.7 urn:nbn:de:0030-drops-134927 https://drops.dagstuhl.de/opus/volltexte/2021/13492/ https://creativecommons.org/licenses/by/3.0/ CC-BY Distributed Machine Learning Single-hop Radio Network Byzantine Fault Communication Complexity Wireless Communication Parameter Server Data processing Computer science InProceedings publishedVersion 2021 ftdagstuhl https://doi.org/10.4230/LIPIcs.OPODIS.2020.7 2022-05-23T15:20:38Z In the past few years, many Byzantine-tolerant distributed machine learning (DML) algorithms have been proposed in the point-to-point communication model. In this paper, we focus on a popular DML framework - the parameter server computation paradigm and iterative learning algorithms that proceed in rounds, e.g., [Gupta and Vaidya, 2020; El-Mhamdi et al., 2020; Chen et al., 2017]. One limitation of prior algorithms in this domain is the high communication complexity. All the Byzantine-tolerant DML algorithms that we are aware of need to send n d-dimensional vectors from worker nodes to the parameter server in each round, where n is the number of workers and d is the number of dimensions of the feature space (which may be in the order of millions). In a wireless network, power consumption is proportional to the number of bits transmitted. Consequently, it is extremely difficult, if not impossible, to deploy these algorithms in power-limited wireless devices. Motivated by this observation, we aim to reduce the communication complexity of Byzantine-tolerant DML algorithms in the single-hop radio network [Alistarh et al., 2010; Bhandari and Vaidya, 2005; Koo, 2004]. Inspired by the CGC filter developed by Gupta and Vaidya, PODC 2020 [Gupta and Vaidya, 2020], we propose a gradient descent-based algorithm, Echo-CGC. Our main novelty is a mechanism to utilize the broadcast properties of the radio network to avoid transmitting the raw gradients (full d-dimensional vectors). In the radio network, each worker is able to overhear previous gradients that were transmitted to the parameter server. Roughly speaking, in Echo-CGC, if a worker "agrees" with a combination of prior gradients, it will broadcast the "echo message" instead of the its raw local gradient. The echo message contains a vector of coefficients (of size at most n) and the ratio of the magnitude between two gradients (a float). In comparison, the traditional approaches need to send n local gradients in each round, where each gradient is typically a vector in a ... Conference Object DML DROPS - Dagstuhl Research Online Publication Server (Schloss Dagstuhl - Leibniz Center for Informatics )
institution Open Polar
collection DROPS - Dagstuhl Research Online Publication Server (Schloss Dagstuhl - Leibniz Center for Informatics )
op_collection_id ftdagstuhl
language English
topic Distributed Machine Learning
Single-hop Radio Network
Byzantine Fault
Communication Complexity
Wireless Communication
Parameter Server
Data processing Computer science
spellingShingle Distributed Machine Learning
Single-hop Radio Network
Byzantine Fault
Communication Complexity
Wireless Communication
Parameter Server
Data processing Computer science
Zhang, Qinzi
Tseng, Lewis
Echo-CGC: A Communication-Efficient Byzantine-Tolerant Distributed Machine Learning Algorithm in Single-Hop Radio Network
topic_facet Distributed Machine Learning
Single-hop Radio Network
Byzantine Fault
Communication Complexity
Wireless Communication
Parameter Server
Data processing Computer science
description In the past few years, many Byzantine-tolerant distributed machine learning (DML) algorithms have been proposed in the point-to-point communication model. In this paper, we focus on a popular DML framework - the parameter server computation paradigm and iterative learning algorithms that proceed in rounds, e.g., [Gupta and Vaidya, 2020; El-Mhamdi et al., 2020; Chen et al., 2017]. One limitation of prior algorithms in this domain is the high communication complexity. All the Byzantine-tolerant DML algorithms that we are aware of need to send n d-dimensional vectors from worker nodes to the parameter server in each round, where n is the number of workers and d is the number of dimensions of the feature space (which may be in the order of millions). In a wireless network, power consumption is proportional to the number of bits transmitted. Consequently, it is extremely difficult, if not impossible, to deploy these algorithms in power-limited wireless devices. Motivated by this observation, we aim to reduce the communication complexity of Byzantine-tolerant DML algorithms in the single-hop radio network [Alistarh et al., 2010; Bhandari and Vaidya, 2005; Koo, 2004]. Inspired by the CGC filter developed by Gupta and Vaidya, PODC 2020 [Gupta and Vaidya, 2020], we propose a gradient descent-based algorithm, Echo-CGC. Our main novelty is a mechanism to utilize the broadcast properties of the radio network to avoid transmitting the raw gradients (full d-dimensional vectors). In the radio network, each worker is able to overhear previous gradients that were transmitted to the parameter server. Roughly speaking, in Echo-CGC, if a worker "agrees" with a combination of prior gradients, it will broadcast the "echo message" instead of the its raw local gradient. The echo message contains a vector of coefficients (of size at most n) and the ratio of the magnitude between two gradients (a float). In comparison, the traditional approaches need to send n local gradients in each round, where each gradient is typically a vector in a ...
format Conference Object
author Zhang, Qinzi
Tseng, Lewis
author_facet Zhang, Qinzi
Tseng, Lewis
author_sort Zhang, Qinzi
title Echo-CGC: A Communication-Efficient Byzantine-Tolerant Distributed Machine Learning Algorithm in Single-Hop Radio Network
title_short Echo-CGC: A Communication-Efficient Byzantine-Tolerant Distributed Machine Learning Algorithm in Single-Hop Radio Network
title_full Echo-CGC: A Communication-Efficient Byzantine-Tolerant Distributed Machine Learning Algorithm in Single-Hop Radio Network
title_fullStr Echo-CGC: A Communication-Efficient Byzantine-Tolerant Distributed Machine Learning Algorithm in Single-Hop Radio Network
title_full_unstemmed Echo-CGC: A Communication-Efficient Byzantine-Tolerant Distributed Machine Learning Algorithm in Single-Hop Radio Network
title_sort echo-cgc: a communication-efficient byzantine-tolerant distributed machine learning algorithm in single-hop radio network
publisher LIPIcs - Leibniz International Proceedings in Informatics. 24th International Conference on Principles of Distributed Systems (OPODIS 2020)
publishDate 2021
url https://doi.org/10.4230/LIPIcs.OPODIS.2020.7
https://nbn-resolving.org/urn:nbn:de:0030-drops-134927
https://drops.dagstuhl.de/opus/volltexte/2021/13492/
genre DML
genre_facet DML
op_relation doi:10.4230/LIPIcs.OPODIS.2020.7
urn:nbn:de:0030-drops-134927
https://drops.dagstuhl.de/opus/volltexte/2021/13492/
op_rights https://creativecommons.org/licenses/by/3.0/
op_rightsnorm CC-BY
op_doi https://doi.org/10.4230/LIPIcs.OPODIS.2020.7
_version_ 1766397362944606208