scispace - formally typeset
Open AccessPosted Content

Gather-Excite: Exploiting Feature Context in Convolutional Neural Networks

Reads0
Chats0
TLDR
Gathering and Excite as mentioned in this paper proposes a pair of operators: gather and excite, which redistributes the pooled information to local features, which can be integrated directly in existing architectures to improve their performance.
Abstract
While the use of bottom-up local operators in convolutional neural networks (CNNs) matches well some of the statistics of natural images, it may also prevent such models from capturing contextual long-range feature interactions. In this work, we propose a simple, lightweight approach for better context exploitation in CNNs. We do so by introducing a pair of operators: gather, which efficiently aggregates feature responses from a large spatial extent, and excite, which redistributes the pooled information to local features. The operators are cheap, both in terms of number of added parameters and computational complexity, and can be integrated directly in existing architectures to improve their performance. Experiments on several datasets show that gather-excite can bring benefits comparable to increasing the depth of a CNN at a fraction of the cost. For example, we find ResNet-50 with gather-excite operators is able to outperform its 101-layer counterpart on ImageNet with no additional learnable parameters. We also propose a parametric gather-excite operator pair which yields further performance gains, relate it to the recently-introduced Squeeze-and-Excitation Networks, and analyse the effects of these changes to the CNN feature activation statistics.

read more

Citations
More filters
Posted Content

Conformer: Local Features Coupling Global Representations for Visual Recognition

TL;DR: Conformer as mentioned in this paper adopts a concurrent structure so that local features and global representations are retained to the maximum extent, and outperforms DeiT-B by 2.3% on ImageNet.
Posted Content

CMT: Convolutional Neural Networks Meet Vision Transformers

TL;DR: CMT-S as discussed by the authors proposes a new transformer based hybrid network by taking advantage of transformers to capture long-range dependencies, and of CNNs to model local features, obtaining much better accuracy and efficiency than previous convolution and transformer based models.
Posted Content

DMSANet: Dual Multi Scale Attention Network.

TL;DR: In this paper, a dual multi-scale attention network (DMSANet) is proposed, where the first part is used to extract features at various scales and aggregate them, the second part uses spatial and channel attention modules in parallel to adaptively integrate local features with their global dependencies.
Posted Content

RaftMLP: Do MLP-based Models Dream of Winning Over Computer Vision?

TL;DR: In this article, a non-convolutional inductive bias was built into the architecture of the MLP-Mixer using two simple ideas: vertically and horizontally dividing the token-mixing block and making spatial correlations denser among some channels of mixing.
Posted Content

Dense xUnit Networks.

TL;DR: This paper adopts and improves the xUnit activation, shows how it can be incorporated into the DenseNet architecture, and illustrates its high effectiveness for classification and image restoration tasks alike.
References
More filters
Proceedings ArticleDOI

Deep Residual Learning for Image Recognition

TL;DR: In this article, the authors proposed a residual learning framework to ease the training of networks that are substantially deeper than those used previously, which won the 1st place on the ILSVRC 2015 classification task.
Proceedings Article

Very Deep Convolutional Networks for Large-Scale Image Recognition

TL;DR: In this paper, the authors investigated the effect of the convolutional network depth on its accuracy in the large-scale image recognition setting and showed that a significant improvement on the prior-art configurations can be achieved by pushing the depth to 16-19 layers.
Journal ArticleDOI

Gradient-based learning applied to document recognition

TL;DR: In this article, a graph transformer network (GTN) is proposed for handwritten character recognition, which can be used to synthesize a complex decision surface that can classify high-dimensional patterns, such as handwritten characters.
Journal ArticleDOI

ImageNet Large Scale Visual Recognition Challenge

TL;DR: The ImageNet Large Scale Visual Recognition Challenge (ILSVRC) as mentioned in this paper is a benchmark in object category classification and detection on hundreds of object categories and millions of images, which has been run annually from 2010 to present, attracting participation from more than fifty institutions.
Posted Content

Faster R-CNN: Towards Real-Time Object Detection with Region Proposal Networks

TL;DR: Faster R-CNN as discussed by the authors proposes a Region Proposal Network (RPN) to generate high-quality region proposals, which are used by Fast R-NN for detection.
Related Papers (5)