Abstract
In-network aggregation (INA) can accelerate distributed deep learning (DDL) by offloading a gradient aggregation operation to programmable switches. However, as the number of jobs increases, the communication overhead increases since the INA switch's resources for aggregation are limited. To reduce the communication overhead, gradient quantization approaches that compress gradients into fewer bits can be used. Although these approaches are effective in INA, existing INA schemes cannot efficiently use the switch's resources for low-bit quantized gradients. In this paper, we propose a quantization-aware in-network aggregation (QAINA) scheme that provides flexible aggregation and achieves efficient resource utilization for low-bit quantized gradients.
Original language | English |
---|---|
Title of host publication | CoNEXT Companion 2023 - Companion of the 19th International Conference on emerging Networking EXperiments and Technologies |
Publisher | Association for Computing Machinery, Inc |
Pages | 57-58 |
Number of pages | 2 |
ISBN (Electronic) | 9798400704079 |
DOIs | |
Publication status | Published - 2023 Dec 5 |
Event | 19th International Conference on emerging Networking EXperiments and Technologies, CoNEXT Companion 2023 - Paris, France Duration: 2023 Dec 5 → 2023 Dec 8 |
Publication series
Name | CoNEXT Companion 2023 - Companion of the 19th International Conference on emerging Networking EXperiments and Technologies |
---|
Conference
Conference | 19th International Conference on emerging Networking EXperiments and Technologies, CoNEXT Companion 2023 |
---|---|
Country/Territory | France |
City | Paris |
Period | 23/12/5 → 23/12/8 |
Bibliographical note
Publisher Copyright:© 2023 Owner/Author.
Keywords
- gradient quantization
- in-network aggregation
ASJC Scopus subject areas
- Computer Networks and Communications
- Computer Science Applications
- Hardware and Architecture