Quantization-Aware In-Network Aggregation for Minimizing Communication Overhead

Hochan Lee, Heewon Kim, Chanbin Bae, Yujin Kim, Sangheon Pack

Research output: Chapter in Book/Report/Conference proceedingConference contribution

1 Citation (Scopus)

Abstract

In-network aggregation (INA) can accelerate distributed deep learning (DDL) by offloading a gradient aggregation operation to programmable switches. However, as the number of jobs increases, the communication overhead increases since the INA switch's resources for aggregation are limited. To reduce the communication overhead, gradient quantization approaches that compress gradients into fewer bits can be used. Although these approaches are effective in INA, existing INA schemes cannot efficiently use the switch's resources for low-bit quantized gradients. In this paper, we propose a quantization-aware in-network aggregation (QAINA) scheme that provides flexible aggregation and achieves efficient resource utilization for low-bit quantized gradients.

Original languageEnglish
Title of host publicationCoNEXT Companion 2023 - Companion of the 19th International Conference on emerging Networking EXperiments and Technologies
PublisherAssociation for Computing Machinery, Inc
Pages57-58
Number of pages2
ISBN (Electronic)9798400704079
DOIs
Publication statusPublished - 2023 Dec 5
Event19th International Conference on emerging Networking EXperiments and Technologies, CoNEXT Companion 2023 - Paris, France
Duration: 2023 Dec 52023 Dec 8

Publication series

NameCoNEXT Companion 2023 - Companion of the 19th International Conference on emerging Networking EXperiments and Technologies

Conference

Conference19th International Conference on emerging Networking EXperiments and Technologies, CoNEXT Companion 2023
Country/TerritoryFrance
CityParis
Period23/12/523/12/8

Bibliographical note

Publisher Copyright:
© 2023 Owner/Author.

Keywords

  • gradient quantization
  • in-network aggregation

ASJC Scopus subject areas

  • Computer Networks and Communications
  • Computer Science Applications
  • Hardware and Architecture

Fingerprint

Dive into the research topics of 'Quantization-Aware In-Network Aggregation for Minimizing Communication Overhead'. Together they form a unique fingerprint.

Cite this