|
|
A Lightweight Convolutional Neural Network Architecture with Slice Feature Map |
ZHANG Yufeng1, ZHENG Zhonglong1, LIU Huawen1, XIANG Daohong2, HE Xiaowei1, LI Zhifei1, HE Yiran1, KHODJA Abd Erraouf1 |
1.Department of Computer Science, Zhejiang Normal University, Jinhua 321004 2.Department of Mathematics, Zhejiang Normal University, Jinhua 321004 |
|
|
Abstract The capacities of mobile and embedded devices are quite inadequate for the requirement of the storage capacity and computational resources of convolutional neural network models. Therefore, a lightweight convolutional neural network architecture, network with slice feature map, named SFNet, is proposed. The concept of slice block is introduced. By performing the “slice” processing on the output feature map of the network, each feature map segment is respectively sent to a convolution kernel of different sizes for convolution operation, and then the obtained feature map is concatenated. A simple 1×1 convolution is utilized to fuse the channels of the feature map. The experiments show that compared with the state-of-the-art lightweight convolutional neural networks, SFNet has fewer parameters and floating-point operations, and higher classification accuracy with the same number of convolution kernels and input feature map channels. Compared with the standard convolution, in the case of a significant reduction in network complexity, the classification accuracy is same or higher.
|
Received: 15 November 2018
|
Fund:Supported by National Natural Science Foundation of China(No.61672467,61572443,11871438) |
Corresponding Authors:
ZHENG Zhonglong. Ph.D., professor. His research inte-rests include pattern recognition, machine learning and image processing.
|
About author:: ZHANG Yufeng, master student. His research interests include deep learning. (LIU Huawen, Ph.D., professor. His research interests include machine learning and data mining.) (XIANG Daohong, Ph.D., associate professor. Her research interests include statistical machine learning, robust statistics and deep learning.) (HE Xiaowei, master, professor. His research interests include machine learning, ima-ge and video processing.) (LI Zhifei, master, lecturer. His research interests include pattern recognition, deep learning and virtual reality.) (HE Yiran, master student. Her research interests include machine learning.)
|
|
|
|
[1] KRIZHEVSKY A, SUTSKEVER I, HINTON G E. ImageNet Classification with Deep Convolutional Neural Networks // PEREIRA F, BURGES C J C, BOTTOU L, et al., eds. Advances in Neural Information Processing Systems 25. Cambridge, USA: The MIT Press, 2012: 1097-1105. [2] REN S Q, HE K M, GIRSHICK R, et al. Faster R-CNN: Towards Real-Time Object Detection with Region Proposal Networks. IEEE Transactions on Pattern Analysis and Machine Intelligence, 2016, 39(6): 1137-1149. [3] LONG J, SHELHAMER E, DARRELL T. Fully Convolutional Networks for Semantic Segmentation // Proc of the IEEE Conference on Computer Vision and Pattern Recognition. Washington, USA: IEEE, 2015: 3431-3440. [4] SIMONYAN K, ZISSERMAN A. Very Deep Convolutional Networks for Large-Scale Image Recognition[C/OL]. [2018-10-24]. https://arxiv.org/pdf/1409.1556.pdf. [5] DENG J, DONG W, SOCHER R, et al. ImageNet: A Large-Scale Hie-rarchical Image Database // Proc of the IEEE Conference on Computer Vision and Pattern Recognition. Washington, USA: IEEE, 2009: 248-255. [6] SZEGEDY C, LIU W, JIA Y Q, et al. Going Deeper with Convolutions // Proc of the IEEE Conference on Computer Vision and Pattern Recognition. Washington, USA: IEEE, 2015. DOI: 10.1109/CVPR.2015.7298594. [7] HE K M, ZHANG X Y, REN S Q, et al. Deep Residual Learning for Image Recognition // Proc of the IEEE Conference on Computer Vision and Pattern Recognition. Washington, USA: IEEE, 2016: 770-778. [8] HUANG G, LIU Z, VAN DER LAURENS M, et al. Densely Connected Convolutional Networks // Proc of the IEEE Conference on Computer Vision and Pattern Recognition. Washington, USA: IEEE, 2017: 2261-2269. [9] HOWARD A G, ZHU M L, CHEN B, et al. MobileNets: Efficient Convolutional Neural Networks for Mobile Vision Applications[C/OL]. [2018-10-24]. https://arxiv.org/pdf/1704.04861.pdf. [10] HAN S, POOL J, TRAN J, et al. Learning Both Weights and Connections for Efficient Neural Network // CORTES C, LAWRENCE N D, LEE D D, et al., eds. Advances in Neural Information Processing Systems 28. Cambridge, USA: The MIT Press, 2015: 1135-1143. [11] NGUYEN H V, ZHOU K, VEMULAPALLI R. Cross-Domain Synthesis of Medical Images Using Efficient Location-Sensitive Deep Network // Proc of the International Conference on Medical Image Computing and Computer-Assisted Intervention. Berlin, Germany: Springer, 2015: 677-684. [12] LI H, KADAV A, DURDANOVIC I, et al. Pruning Filters for Efficient ConvNets[C/OL]. [2018-10-24]. https://arxiv.org/pdf/1608.08710.pdf. [13] HAN S, MAO H Z, DALLY W J. Deep Compression: Compre-ssing Deep Neural Networks with Pruning, Trained Quantization and Huffman Coding[C/OL]. [2018-10-24]. https://arxiv.org/pdf/1510.00149.pdf. [14] CHEN W L, WILSON J T, TYREE S, et al. Compressing Neural Networks with the Hashing Trick // Proc of the 32nd International Conference on Machine Learning. Berlin, Germany: Springer, 2015: 2285-2294. [15] DENTON E, ZAREMBA W, BRUNA J, et al. Exploiting Linear Structure within Convolutional Networks for Efficient Evaluation // Proc of the 27th International Conference on Neural Information Processing Systems. Cambridge, USA: The MIT Press, 2014: 1269-1277. [16] SIRONI A, TEKIN B, RIGAMONTI R, et al. Learning Separable Filters. IEEE Transactions on Pattern Analysis and Machine Intelligence, 2015, 37(1): 94-116. [17] JADERBERG M, VEDALDI A, ZISSERMAN A. Speeding up Convolutional Neural Networks with Low Rank Expansions[C/OL]. [2018-10-24]. https://arxiv.org/pdf/1405.3866.pdf. [18] SOTOUDEH M, BAGHSORKHI S S. DeepThin: A Self-Compre-ssing Library for Deep Neural Networks[C/OL]. [2018-10-24]. https://arxiv.org/pdf/1802.06944.pdf. [19] VANHOUCKE V, SENIOR A, MAO M Z. Improving the Speed of Neural Networks on CPUs // Proc of the NIPS Workshop on Deep Learning and Unsupervised Feature Learning. Berlin, Germany: Springer, 2011, I: 611-620. [20] ARORA S, BHASKARA A, GE R, et al. Provable Bounds for Learning Some Deep Representations[C/OL]. [2018-10-24]. https://arxiv.org/pdf/1310.6343.pdf. [21] HWANG K, SUNG W. Fixed-Point Feedforward Deep Neural Network Design Using Weights +1, 0, and -1 // Proc of the IEEE Workshop on Signal Processing Systems. Washington, USA: IEEE, 2014. DOI: 10.1109/SiPS.2014.6986082. [22] COURBARIAUX M, BENGIO Y, DAVID J P. BinaryConnect: Training Deep Neural Networks with Binary Weights During Propagations // CORTES C, LAWRENCE N D, LEE D D, et al., eds. Advances in Neural Information Processing Systems 28. Cambridge, USA: The MIT Press, 2015: 3123-3131. [23] COURBARIAUX M, BENGIO Y. Binarized Neural Networks: Training Deep Neural Networks with Weights and Activations Constrained to +1 or -1[C/OL]. [2018-10-24]. https://arxiv.org/pdf/1602.02830.pdf. [24] RASTEGAN M, ORDONEZ V, REDMON J, et al. XNOR-Net: ImageNet Classification Using Binary Convolutional Neural Networks // Proc of the European Conference on Computer Vision. Berlin, Germany: Springer, 2016: 525-542. [25] BUCILUA‘ C, CARUANA R, NICULESCU-MIZIL A. Model Compression // Proc of the 12th ACM SIGKDD International Confe-rence on Knowledge Discovery and Data Mining. New York, USA: ACM, 2006: 535-541. [26] BA L J, CARUANA R. Do Deep Nets Really Need to be Deep? [C/OL]. [2018-10-24]. https://arxiv.org/pdf/1312.6184v5.pdf. [27] HINTON G, VINYALS O, DEAN J. Distilling the Knowledge in a Neural Network[C/OL]. [2018-10-24]. https://arxiv.org/pdf/1503.02531.pdf. [28] ZEILER M D, FERGUS R. Visualizing and Understanding Convolutional Networks // Proc of the European Conference on Computer Vision. Berlin, Germany: Springer, 2013: 818-833. [29] SZEGEDY C, VANHOUCKE V, IOFFE S, et al. Rethinking the Inception Architecture for Computer Vision // Proc of the IEEE Conference on Computer Vision and Pattern Recognition. Washington, USA: IEEE, 2016: 2818-2826. [30] LIN M, CHEN Q, YAN S C. Network in Network[C/OL]. [2018-10-24]. https://arxiv.org/pdf/1312.4400.pdf. [31] SZEGEDY C, IOFFE S, VANHOUCKE V, et al. Inception-v4, Inception-ResNet and the Impact of Residual Connections on Learning[C/OL]. [2018-10-24]. https://arxiv.org/pdf/1602.07261.pdf. [32] IANDOLA F N, HAN S, MOSKEWICZ M W, et al. SqueezeNet: AlexNet-Level Accuracy with 50x Fewer Parameters and <0.5MB Model Size[C/OL]. [2018-10-24]. https://arxiv.org/pdf/1602.07360.pdf. [33] IOANNOU Y, ROBERTSON D, SHOTTON J, et al. Training CN-Ns with Low-Rank Filters for Efficient Image Classification[C/OL]. [2018-10-24]. https://arxiv.org/pdf/1511.06744.pdf. [34] IOANNOU Y, ROBERTSON D, CIPOLLA R, et al. Deep Roots: Improving CNN Efficiency with Hierarchical Filter Groups // Proc of the IEEE Conference on Computer Vision and Pattern Recognition. Washington, USA: IEEE, 2017: 5977-5986. [35] XIE S N, GIRSHICK R, DOLLAR P, et al. Aggregated Residual Transformations for Deep Neural Networks // Proc of the IEEE Conference on Computer Vision and Pattern Recognition. Washington, USA: IEEE, 2017: 5987-5995. [36] CHOLLET F. Xception: Deep Learning with Depthwise Separable Convolutions // Proc of the IEEE Conference on Computer Vision and Pattern Recognition. Washington, USA: IEEE, 2017: 1800-1807. [37] ZHANG X Y, ZHOU X Y, LIN M X, et al. ShuffleNet: An Extremely Efficient Convolutional Neural Network for Mobile Devices[C/OL]. [2018-10-24]. https://arxiv.org/pdf/1707.01083.pdf. [38] ZHANG T, QI G J, XIAO B, et al. Interleaved Group Convolutions for Deep Neural Networks // Proc of the IEEE International Conference on Computer Vision. Washington, USA: IEEE, 2017: 4383-4392. [39] SANDLER M, HOWARD A, ZHU M L, et al. MobileNetV2: Inverted Residuals and Linear Bottlenecks[C/OL]. [2018-10-24]. https://arxiv.org/pdf/1801.04381v3.pdf. [40] IOFFE S, SZEGEDY C. Batch Normalization: Accelerating Deep Network Training by Reducing Internal Covariate Shift // Proc of the 32nd International Conference on Machine Learning. New York, USA: Springer, 2015: 448-456. [41] JIA Y Q, SHELHAMER E, DONAHUE J, et al. Caffe: Convolutional Architecture for Fast Feature Embedding[C/OL]. [2018-10-24]. https://arxiv.org/pdf/1408.5093.pdf. [42] KRIZHEVSKY A. Learning Multiple Layers of Features from Tiny Images[C/OL]. [2018-10-24]. https://www.cs.toronto.edu/~kriz/learning-features-2009-TR.pdf. [43] NETZER Y, WANG T, COATES A, et al. Reading Digits in Natural Images with Unsupervised Feature Learning // Proc of the NIPS Workshop on Deep Learning and Unsupervised Feature Learning. Berlin, Germany: Springer, 2011. DOI: 10.2118/18761-MS. [44] FREEMAN I, ROESE-KOERNER L, KUMMERT A. Effnet: An Efficient Structure For Convolutional Neural Networks // Proc of the 25th IEEE International Conference on Image Processing. Washington, USA: IEEE, 2018: 6-10. [45] EVERINGHAM M, VAN GOOL L, WILLIAMS C K I, et al. The PASCAL Visual Object Classes(VOC) Challenge. International Journal of Computer Vision, 2010, 88(2): 303-338.. [46] LIU W, ANGUELOV D, ERHAN D, et al.SSD: Single Shot Mul-tiBox Detector // Proc of the European Conference on Computer Vision. Berlin, Germany: Springer, 2016: 21-37. |
|
|
|