ETRI-Knowledge Sharing Plaform

KOREAN
논문 검색
Type SCI
Year ~ Keyword

Detail

Conference Paper CPrune: Compiler-Informed Model Pruning for Efficient Target-Aware DNN Execution
Cited 2 time in scopus Download 19 time Share share facebook twitter linkedin kakaostory
Authors
Taeho Kim, Yongin Kwon, Jemin Lee, Taeho Kim, Sangtae Ha
Issue Date
2022-10
Citation
European Conference on Computer Vision (ECCV) 2022 (LNCS 13680), pp.651-667
Language
English
Type
Conference Paper
DOI
https://dx.doi.org/10.1007/978-3-031-20044-1_37
Project Code
22HS2300, Neuromorphic Computing Software Platform for Artificial Intelligence Systems, Taeho Kim
Abstract
Mobile devices run deep learning models for various purposes, such as image classification and speech recognition. Due to the resource constraints of mobile devices, researchers have focused on either making a lightweight deep neural network (DNN) model using model pruning or generating an efficient code using compiler optimization. Surprisingly, we found that the straightforward integration between model compression and compiler auto-tuning often does not produce the most efficient model for a target device. We propose CPrune, a compiler-informed model pruning for efficient target-aware DNN execution to support an application with a required target accuracy. CPrune makes a lightweight DNN model through informed pruning based on the structural information of subgraphs built during the compiler tuning process. Our experimental results show that CPrune increases the DNN execution speed up to 2.73 × compared to the state-of-the-art TVM auto-tune while satisfying the accuracy requirement.
KSP Keywords
Auto-tune, Deep neural network(DNN), Execution speed, Image classification, Mobile devices, Model compression, Resource constraints, Speed-up, Structural information, Target accuracy, Tuning process