Privacy-Sensitive Parallel Split Learning

Research output: Chapter in Book/Report/Conference proceedingConference contribution

Abstract

Mobile devices and medical centers have access to rich data that is suitable for training deep learning models. However, these highly distributed datasets are privacy sensitive making privacy issues for applying deep learning techniques to the problem at hand. Split Learning can solve these data privacy problems, but the possibility of overfitting exists because each node doesn't train in parallel but in a sequential manner. In this paper, we propose a parallel split learning method that prevents overfitting due to differences in a training order and data size by the node. Our method selects mini-batch size considering the amount of local data on each node and synchronizes the layers that nodes have during the training process so that all nodes can use the equivalent deep learning model when the training is complete.

Original languageEnglish
Title of host publication34th International Conference on Information Networking, ICOIN 2020
PublisherIEEE Computer Society
Pages7-9
Number of pages3
ISBN (Electronic)9781728141985
DOIs
Publication statusPublished - 2020 Jan
Event34th International Conference on Information Networking, ICOIN 2020 - Barcelona, Spain
Duration: 2020 Jan 72020 Jan 10

Publication series

NameInternational Conference on Information Networking
Volume2020-January
ISSN (Print)1976-7684

Conference

Conference34th International Conference on Information Networking, ICOIN 2020
Country/TerritorySpain
CityBarcelona
Period20/1/720/1/10

Bibliographical note

Publisher Copyright:
© 2020 IEEE.

Keywords

  • Distributed Deep Learning
  • Federated Learning
  • Split Learning

ASJC Scopus subject areas

  • Computer Networks and Communications
  • Information Systems

Fingerprint

Dive into the research topics of 'Privacy-Sensitive Parallel Split Learning'. Together they form a unique fingerprint.

Cite this