Abstract
One of the reasons behind the tremendous success of deep learning theory and applications in the recent days is advances in distributed and parallel high performance computing (HPC). This paper proposes a new distributed deep learning platform, named ShmCaffe, which utilizes remote shared memory for communication overhead reduction in massive deep neural network training parameter sharing. ShmCaffe is designed based on Soft Memory Box (SMB), a virtual shared memory framework. In the SMB framework, the remote shared memory is used as a shared buffer for asynchronous massive parameter sharing among many distributed deep learning processes. Moreover, a hybrid method that combines asynchronous and synchronous parameter sharing methods is also discussed in this paper for improving scalability. As a result, ShmCaffe is 10.1 times faster than Caffe and 2.8 times faster than Caffe-MPI for deep neural network training when Inception\-v1 is trained with 16 GPUs. We verify the convergence of the Inception\-v1 model training using ShmCaffe-A and ShmCaffe-H by varying the number of workers. Furthermore, we evaluate scalability of ShmCaffe by analyzing the computation and communication times per one iteration of deep learning training in four convolutional neural network (CNN) models.
| Original language | English |
|---|---|
| Title of host publication | Proceedings - 2018 IEEE 38th International Conference on Distributed Computing Systems, ICDCS 2018 |
| Publisher | Institute of Electrical and Electronics Engineers Inc. |
| Pages | 1118-1128 |
| Number of pages | 11 |
| ISBN (Electronic) | 9781538668719 |
| DOIs | |
| Publication status | Published - 2018 Jul 19 |
| Externally published | Yes |
| Event | 38th IEEE International Conference on Distributed Computing Systems, ICDCS 2018 - Vienna, Austria Duration: 2018 Jul 2 → 2018 Jul 5 |
Publication series
| Name | Proceedings - International Conference on Distributed Computing Systems |
|---|---|
| Volume | 2018-July |
Conference
| Conference | 38th IEEE International Conference on Distributed Computing Systems, ICDCS 2018 |
|---|---|
| Country/Territory | Austria |
| City | Vienna |
| Period | 18/7/2 → 18/7/5 |
Bibliographical note
Publisher Copyright:© 2018 IEEE.
Keywords
- Deep learning
- Distributed deep learning
- Shared memory
- ShmCaffe
- Soft memory box
ASJC Scopus subject areas
- Software
- Hardware and Architecture
- Computer Networks and Communications