Squeeze-and-Excitation (SE) block presents a channel attention mechanism for modeling global context via explicitly capturing dependencies across channels. However, we are still far from understanding how the SE block works. In this work, we first revisit the SE block, and then present a detailed empirical study of the relationship between global context and attention distribution, based on which we propose a simple yet effective module, called Linear Context Transform (LCT) block. We divide all channels into different groups and normalize the globally aggregated context features within each channel group, reducing the disturbance from irrelevant channels. Through linear transform of the normalized context features, we model global context ...
The capability of the self-attention mechanism to model the long-range dependencies has catapulted i...
Abstract Object detection is an important component of computer vision. Most of the recent successfu...
Recently, attention mechanisms have been explored with ConvNets, both across the spatial and channel...
Squeeze-and-Excitation (SE) block presents a channel attention mechanism for modeling global context...
While the use of bottom-up local operators in convolutional neural networks (CNNs) matches well some...
Self-attention mechanisms model long-range context by using pairwise attention between all input tok...
Entropy modeling is a key component for high-performance image compression algorithms. Recent develo...
We propose global context vision transformer (GC ViT), a novel architecture that enhances parameter ...
In this paper, we propose a general framework for image classification using the attention mechanism...
Pursuing an object detector with good detection accuracy while ensuring detection speed has always b...
In this paper, we aim to enhance self-attention (SA) mechanism for deep metric learning in visual pe...
In this work, we propose a generally applicable transformation unit for visual recognition with deep...
In the recent advancements attention mechanism in deep learning had played a vital role in proving b...
Semantic segmentation is a task that covers most of the perception needs of intelligent vehicles in ...
Attention mechanisms have demonstrated great potential in improving the performance of deep convolut...
The capability of the self-attention mechanism to model the long-range dependencies has catapulted i...
Abstract Object detection is an important component of computer vision. Most of the recent successfu...
Recently, attention mechanisms have been explored with ConvNets, both across the spatial and channel...
Squeeze-and-Excitation (SE) block presents a channel attention mechanism for modeling global context...
While the use of bottom-up local operators in convolutional neural networks (CNNs) matches well some...
Self-attention mechanisms model long-range context by using pairwise attention between all input tok...
Entropy modeling is a key component for high-performance image compression algorithms. Recent develo...
We propose global context vision transformer (GC ViT), a novel architecture that enhances parameter ...
In this paper, we propose a general framework for image classification using the attention mechanism...
Pursuing an object detector with good detection accuracy while ensuring detection speed has always b...
In this paper, we aim to enhance self-attention (SA) mechanism for deep metric learning in visual pe...
In this work, we propose a generally applicable transformation unit for visual recognition with deep...
In the recent advancements attention mechanism in deep learning had played a vital role in proving b...
Semantic segmentation is a task that covers most of the perception needs of intelligent vehicles in ...
Attention mechanisms have demonstrated great potential in improving the performance of deep convolut...
The capability of the self-attention mechanism to model the long-range dependencies has catapulted i...
Abstract Object detection is an important component of computer vision. Most of the recent successfu...
Recently, attention mechanisms have been explored with ConvNets, both across the spatial and channel...