LightweightNet: Toward fast and lightweight convolutional neural networks via architecture distillation
作者:
Highlights:
• We present a new framework of deep convolutional neural network architecture distillation, namely LightweightNet, for acceleration and compression.
• We exploit the prior knowledge of pre-defined network architecture to guide the efficient design of acceleration/compression strategies, while not using pre-trained model.
• The proposed framework consists of network parameter compression, network structure acceleration, and non-tensor layer improvement.
• The proposed framework demonstrates a higher acceleration/compression rate than previous methods in experiments, including a large category handwritten Chinese character recognition task with state-of-the-art performance.
摘要
•We present a new framework of deep convolutional neural network architecture distillation, namely LightweightNet, for acceleration and compression.•We exploit the prior knowledge of pre-defined network architecture to guide the efficient design of acceleration/compression strategies, while not using pre-trained model.•The proposed framework consists of network parameter compression, network structure acceleration, and non-tensor layer improvement.•The proposed framework demonstrates a higher acceleration/compression rate than previous methods in experiments, including a large category handwritten Chinese character recognition task with state-of-the-art performance.
论文关键词:Deep network acceleration and compression,Architecture distillation,Lightweight network
论文评审过程:Received 11 February 2018, Revised 4 July 2018, Accepted 28 October 2018, Available online 2 November 2018, Version of Record 27 November 2018.
论文官网地址:https://doi.org/10.1016/j.patcog.2018.10.029