Abstract
A majority of training algorithms for deep neural networks (DNNs) use variants of stochastic gradient descent (SGD). Since a large amount of time is typically required to train DNNs, many attempts have been made to speed-up the training process by parallelizing the SGD algorithms. However, such parallelization efforts introduce approximation due to the inherent sequential nature of the SGD methods. In this paper, we revisit and analyze parallel SGD algorithms, and propose a novel pipelined SGD that is more efficient than previous algorithms.
Original language | English |
---|---|
Title of host publication | Proceedings - 2018 International Conference on Computational Science and Computational Intelligence, CSCI 2018 |
Publisher | Institute of Electrical and Electronics Engineers Inc. |
Pages | 1462-1463 |
Number of pages | 2 |
ISBN (Electronic) | 9781728113609 |
DOIs | |
Publication status | Published - 2018 Dec |
Event | 2018 International Conference on Computational Science and Computational Intelligence, CSCI 2018 - Las Vegas, United States Duration: 2018 Dec 13 → 2018 Dec 15 |
Publication series
Name | Proceedings - 2018 International Conference on Computational Science and Computational Intelligence, CSCI 2018 |
---|
Conference
Conference | 2018 International Conference on Computational Science and Computational Intelligence, CSCI 2018 |
---|---|
Country/Territory | United States |
City | Las Vegas |
Period | 18/12/13 → 18/12/15 |
Bibliographical note
Funding Information:ACKNOWLEDGMENT This work was supported by Basic Science Research Program through the National Research Foundation of Korea (NRF) funded by the Ministry of Science, ICT and Future Planning (NRF-2017R1E1A1A01078157). Also, it was partly supported by the Institute for Information & Communications Technology Promotion (IITP) grant funded by the Korean government (MSIP) (No.2018-0-00269, A research on safe and convenient big data processing methods).
Publisher Copyright:
© 2018 IEEE.
Keywords
- Deep neural networks
- Pipelined stochastic gradient descent
- Stochastic gradient descent
ASJC Scopus subject areas
- Computer Networks and Communications
- Computer Science Applications
- Hardware and Architecture
- Information Systems and Management
- Control and Optimization
- Modelling and Simulation
- Artificial Intelligence