ETRI-Knowledge Sharing Plaform

KOREAN
논문 검색
Type SCI
Year ~ Keyword

Detail

Conference Paper Distributed Deep Learning Framework based on Shared Memory for Fast Deep Neural Network Training
Cited 12 time in scopus Share share facebook twitter linkedin kakaostory
Authors
Eun-Ji Lim, Shin-Young Ahn, Yoo-Mi Park, Wan Choi
Issue Date
2018-10
Citation
International Conference on Information and Communication Technology Convergence (ICTC) 2018, pp.1239-1242
Publisher
IEEE
Language
English
Type
Conference Paper
DOI
https://dx.doi.org/10.1109/ICTC.2018.8539420
Abstract
In distributed deep neural network training, since the communication overhead caused by parameter sharing across multiple deep learning workers can be a performance bottleneck, performing efficient parameter sharing is a crucial challenge in distributed deep learning framework. In this paper, we propose a distributed deep learning framework called TFSM, uses remote shared memory for efficient parameter sharing to accelerate distributed DNN training. TFSM is based on the remote shared memory framework which provides shared memory accessible by multi-machines at high-speed. TFSM provides a new asynchronous parameter update method based on the remote shared memory. We confirmed that the TFSM improves the training time of DNN compared to TensorFlow by training well-known deep learning models using 8 GPU workers.
KSP Keywords
Communication overhead, Deep learning framework, Deep neural network(DNN), High Speed, Neural network training, Shared Memory, Training time, deep learning(DL), deep learning models, parameter sharing, performance bottleneck