Amc: Automl for model compression and acceleration on mobile devices

Y He, J Lin, Z Liu, H Wang, LJ Li… - Proceedings of the …, 2018 - openaccess.thecvf.com
Proceedings of the European conference on computer vision (ECCV), 2018openaccess.thecvf.com
Abstract Model compression is an effective technique to efficiently deploy neural network
models on mobile devices which have limited computation resources and tight power
budgets. Conventional model compression techniques rely on hand-crafted features and
require domain experts to explore the large design space trading off among model size,
speed, and accuracy, which is usually sub-optimal and time-consuming. In this paper, we
propose AutoML for Model Compression (AMC) which leverages reinforcement learning to …
Abstract
Model compression is an effective technique to efficiently deploy neural network models on mobile devices which have limited computation resources and tight power budgets. Conventional model compression techniques rely on hand-crafted features and require domain experts to explore the large design space trading off among model size, speed, and accuracy, which is usually sub-optimal and time-consuming. In this paper, we propose AutoML for Model Compression (AMC) which leverages reinforcement learning to efficiently sample the design space and can improve the model compression quality. We achieved state-of-the-art model compression results in a fully automated way without any human efforts. Under 4× FLOPs reduction, we achieved 2.7% better accuracy than the hand-crafted model compression method for VGG-16 on ImageNet. We applied this automated, push-the-button compression pipeline to MobileNet-V1 and achieved a speedup of 1.53 x on the GPU (Titan Xp) and 1.95 x on an Android phone (Google Pixel 1), with negligible loss of accuracy.
openaccess.thecvf.com
以上显示的是最相近的搜索结果。 查看全部搜索结果