ETRI-Knowledge Sharing Plaform

ENGLISH

성과물

논문 검색
구분 SCI
연도 ~ 키워드

상세정보

학술대회 Improved Early Exiting Activation to Accelerate Edge Inference
Cited 2 time in scopus Download 8 time Share share facebook twitter linkedin kakaostory
저자
박준용, 이종률, 문용혁
발행일
202110
출처
International Conference on Information and Communication Technology Convergence (ICTC) 2021, pp.1813-1817
DOI
https://dx.doi.org/10.1109/ICTC52510.2021.9621109
협약과제
21HS7200, 능동적 즉시 대응 및 빠른 학습이 가능한 적응형 경량 엣지 연동분석 기술개발, 문용혁
초록
As mobile edge devices are getting powerful, on-device deep learning is becoming a reality. However, there are still many challenges for deep learning edge inferences, such as limited resources such as computing power, memory space, and energy. To address these challenges, model compression such as channel pruning, low rank representation, network quantization, and early exiting has been introduce to reduce the computational load of neural networks at a whole. In this paper, we propose an improved method of implementing early exiting branches on a pre-defined neural network, so that it can determine whether the input data is easy to process, therefore use less resource to execute the task. Our method starts with an entire search for activations in a given network, then inserting early exiting modules, testing those early exit branches, resulting in selecting useful branches that are both accurate and fast. Our contribution is reducing the computing time of neural networks by breaking the flow of models using execution branches. Additionally, by testing on all activations in neural network, we gain knowledge of the neural network model and insight on where to place the ideal early exit auxiliary classifier. We test on ResNet model and show reduction in real computation time on single input images.
KSP 제안 키워드
Computing power, Computing time, Edge devices, Improved method, Limited resources, Low rank representation, Memory space, Model compression, Single-input, computation time, computational load