ETRI-Knowledge Sharing Plaform

ENGLISH

성과물

논문 검색
구분 SCI
연도 ~ 키워드

상세정보

학술지 Hardware Resource Analysis in Distributed Training with Edge Devices
Cited 2 time in scopus Download 22 time Share share facebook twitter linkedin kakaostory
저자
박시형, 이제민, 김형신
발행일
202001
출처
Electronics, v.9 no.1, pp.1-13
ISSN
2079-9292
출판사
MDPI
DOI
https://dx.doi.org/10.3390/electronics9010028
협약과제
19HS4800, 인공지능 시스템을 위한 뉴로모픽 컴퓨팅 SW 플랫폼 기술 개발, 김태호
초록
When training a deep learning model with distributed training, the hardware resource utilization of each device depends on the model structure and the number of devices used for training. Distributed training has recently been applied to edge computing. Since edge devices have hardware resource limitations such as memory, there is a need for training methods that use hardware resources efficiently. Previous research focused on reducing training time by optimizing the synchronization process between edge devices or by compressing the models. In this paper, we monitored hardware resource usage based on the number of layers and the batch size of the model during distributed training with edge devices. We analyzed memory usage and training time variability as the batch size and number of layers increased. Experimental results demonstrated that, the larger the batch size, the fewer synchronizations between devices, resulting in less accurate training. In the shallow model, training time increased as the number of devices used for training increased because the synchronization between devices took more time than the computation time of training. This paper finds that efficient use of hardware resources for distributed training requires selecting devices in the context of model complexity and that fewer layers and smaller batches are required for efficient hardware use.
키워드
Deep learning, Distributed training, Edge computing, Internet of Things, Performance monitoring
KSP 제안 키워드
Batch size, Distributed training, Edge devices, Hardware Resources, Internet of thing(IoT), Learning model, Model structure, Number of layers, Performance monitoring, Resource usage, Training time
본 저작물은 크리에이티브 커먼즈 저작자 표시 (CC BY) 조건에 따라 이용할 수 있습니다.
저작자 표시 (CC BY)