Content-aware convolutional neural networks.

Yong Guo, Yaofo Chen, Mingkui Tan, Kui Jia, Jian Chen, Jingdong Wang
Author Information
  1. Yong Guo: South China University of Technology, China; Pazhou Laboratory, China; Key Laboratory of Big Data and Intelligent Robot, Ministry of Education. Electronic address: guo.yong@mail.scut.edu.cn.
  2. Yaofo Chen: South China University of Technology, China. Electronic address: sechenyaofo@mail.scut.edu.cn.
  3. Mingkui Tan: South China University of Technology, China; Key Laboratory of Big Data and Intelligent Robot, Ministry of Education. Electronic address: mingkuitan@scut.edu.cn.
  4. Kui Jia: South China University of Technology, China. Electronic address: kuijia@scut.edu.cn.
  5. Jian Chen: South China University of Technology, China. Electronic address: ellachen@scut.edu.cn.
  6. Jingdong Wang: Microsoft Research Asia, China. Electronic address: jingdw@microsoft.com.

Abstract

Convolutional Neural Networks (CNNs) have achieved great success due to the powerful feature learning ability of convolution layers. Specifically, the standard convolution traverses the input images/features using a sliding window scheme to extract features. However, not all the windows contribute equally to the prediction results of CNNs. In practice, the convolutional operation on some of the windows (e.g., smooth windows that contain very similar pixels) can be very redundant and may introduce noises into the computation. Such redundancy may not only deteriorate the performance but also incur the unnecessary computational cost. Thus, it is important to reduce the computational redundancy of convolution to improve the performance. To this end, we propose a Content-aware Convolution (CAC) that automatically detects the smooth windows and applies a 1 ×1 convolutional kernel to replace the original large kernel. In this sense, we are able to effectively avoid the redundant computation on similar pixels. By replacing the standard convolution in CNNs with our CAC, the resultant models yield significantly better performance and lower computational cost than the baseline models with the standard convolution. More critically, we are able to dynamically allocate suitable computation resources according to the data smoothness of different images, making it possible for content-aware computation. Extensive experiments on various computer vision tasks demonstrate the superiority of our method over existing methods.

Keywords

MeSH Term

Algorithms
Neural Networks, Computer

Word Cloud

Created with Highcharts 10.0.0convolutionwindowscomputationCNNsstandardconvolutionalperformancecomputationalNeuralsmoothsimilarpixelsredundantmayredundancycostContent-awareConvolutionCACkernelablemodelsnetworksConvolutionalNetworksachievedgreatsuccessduepowerfulfeaturelearningabilitylayersSpecificallytraversesinputimages/featuresusingslidingwindowschemeextractfeaturesHowevercontributeequallypredictionresultspracticeoperationegcontaincanintroducenoisesdeterioratealsoincurunnecessaryThusimportantreduceimproveendproposeautomaticallydetectsapplies1×1replaceoriginallargesenseeffectivelyavoidreplacingresultantyieldsignificantlybetterlowerbaselinecriticallydynamicallyallocatesuitableresourcesaccordingdatasmoothnessdifferentimagesmakingpossiblecontent-awareExtensiveexperimentsvariouscomputervisiontasksdemonstratesuperioritymethodexistingmethodsneuralRedundancyreduction

Similar Articles

Cited By