陈志奎

个人信息Personal Information

教授

博士生导师

硕士生导师

主要任职:teaching

性别:男

毕业院校:重庆大学

学位:博士

所在单位:软件学院、国际信息与软件学院

学科:软件工程. 计算机软件与理论

办公地点:开发区综合楼405

联系方式:Email: zkchen@dlut.edu.cn Moble:13478461921 微信:13478461921 QQ:1062258606

电子邮箱:zkchen@dlut.edu.cn

扫描关注

论文成果

当前位置: 中文主页 >> 科学研究 >> 论文成果

Cross-Entropy Pruning for Compressing Convolutional Neural Networks

点击次数:

论文类型:期刊论文

发表时间:2018-11-01

发表刊物:NEURAL COMPUTATION

收录刊物:PubMed、SCIE、Scopus

卷号:30

期号:11

页面范围:3128-3149

ISSN号:0899-7667

关键字:Digital storage; Errors; Neural networks; Object detection, Convolutional neural network; Cross entropy; Data set; Output neurons; Sparse modeling; Storage costs, Entropy

摘要:The success of CNNs is accompanied by deep models and heavy storage costs. For compressing CNNs, we propose an efficient and robust pruning approach, cross-entropy pruning (CEP). Given a trained CNN model, connections were divided into groups in a group-wise way according to their corresponding output neurons. All connections with their cross-entropy errors below a grouping threshold were then removed. A sparse model was obtained and the number of parameters in the baseline model significantly reduced. This letter also presents a highest cross-entropy pruning (HCEP) method that keeps a small portion of weights with the highest CEP. This method further improves the accuracy of CEP. To validate CEP, we conducted the experiments on low redundant networks that are hard to compress. For the MNIST data set, CEP achieves an 0.08% accuracy drop required by LeNet-5 benchmark with only 16% of original parameters. Our proposed CEP also reduces approximately 75% of the storage cost of AlexNet on the ILSVRC 2012 data set, increasing the top-1 errorby only 0.4% and top-5 error by only 0.2%. Compared with three existing methods on LeNet-5, our proposed CEP and HCEP perform significantly better than the existing methods in terms of the accuracy and stability. Some computer vision tasks on CNNs such as object detection and style transfer can be computed in a high-performance way using our CEP and HCEP strategies.