Abstract
Convolution operation in deep convolutional neural networks is the most computationally expensive as compared to other operations. Most of the model computation (FLOPS) in the deep architecture belong to convolution operation. In this paper, we are proposing a novel skip convolution operation that employs significantly fewer computation as compared to the traditional one without sacrificing model accuracy. Skip convolution operation produces structured sparsity in the output feature maps without requiring sparsity in the model parameters for computation reduction. The existing convolution operation performs the redundant computation for object feature representation while the proposed convolution skips redundant computation. Our empirical evaluation for various deep models (VGG, ResNet, MobileNet, and Faster R-CNN) over various benchmarked datasets (CIFAR-10, CIFAR-100, ImageNet, and MS-COCO) show that skip convolution reduces the computation significantly while preserving feature representational capacity. The proposed approach is model-agnostic and can be applied over any architecture. The proposed approach does not require a pretrained model and does train from scratch. Hence we achieve significant computation reduction at training and test time. We are also able to reduce computation in an already compact model such as MobileNet using skip convolution. We also show empirically that the proposed convolution works well for other tasks such as object detection. Therefore, SkipConv can be a widely usable and efficient way of reducing computation in deep CNN models.
Original language | English |
---|---|
Title of host publication | 2020 International Joint Conference on Neural Networks, IJCNN 2020 - Proceedings |
Publisher | IEEE |
ISBN (Electronic) | 9781728169262 |
DOIs | |
Publication status | Published - 28 Sept 2020 |
Event | 2020 International Joint Conference on Neural Networks, IJCNN 2020 - Virtual, Glasgow, UK United Kingdom Duration: 19 Jul 2020 → 24 Jul 2020 |
Publication series
Name | Proceedings of the International Joint Conference on Neural Networks |
---|---|
Volume | 2020 |
ISSN (Print) | 2161-4393 |
ISSN (Electronic) | 2161-4407 |
Conference
Conference | 2020 International Joint Conference on Neural Networks, IJCNN 2020 |
---|---|
Country/Territory | UK United Kingdom |
City | Virtual, Glasgow |
Period | 19/07/20 → 24/07/20 |
Bibliographical note
Publisher Copyright:© 2020 IEEE.
Copyright:
Copyright 2020 Elsevier B.V., All rights reserved.
Keywords
- Computation (FLOPS) compression
- Computationally efficient CNN
- Convolutional neural network
- Deep learning
- Image recognition
ASJC Scopus subject areas
- Software
- Artificial Intelligence