Quantization is one of the most effective methods to compress neural networks, which has achieved great success on convolutional neural networks (CNNs). Recently, vision transformers have demonstrated great potential in computer vision. However, previous post-training quantization methods performed not well on vision transformer, resulting in more than 1% accuracy drop even in 8-bit quantization. Therefore, we analyze the problems of quantization on vision transformers. We observe the distributions of activation values after softmax and GELU functions are quite different from the Gaussian distribution. We also observe that common quantization metrics, such as MSE and cosine distance, are inaccurate to determine the optimal scaling factor. I...
Vision transformers (ViTs) have recently obtained success in many applications, but their intensive ...
Vision Transformers (ViT) have been shown to attain highly competitive performance for a wide range ...
The great success of transformer-based models in natural language processing (NLP) has led to variou...
Network quantization significantly reduces model inference complexity and has been widely used in re...
In this paper, we propose a fully differentiable quantization method for vision transformer (ViT) na...
Vision transformers have recently gained great success on various computer vision tasks; nevertheles...
Data-free quantization can potentially address data privacy and security concerns in model compressi...
Vision Transformer (ViT) architectures are becoming increasingly popular and widely employed to tack...
Attention-based neural networks such as the Vision Transformer (ViT) have recently attained state-of...
Vision Transformers (ViTs) have achieved state-of-the-art performance on various computer vision app...
Vision transformers (ViTs) have recently obtained success in many applications, but their intensive ...
While neural networks have been remarkably successful in a wide array of applications, implementing ...
The neural network quantization is highly desired procedure to perform before running neural network...
We attempt to reduce the computational costs in vision transformers (ViTs), which increase quadratic...
The vision transformer (ViT) has advanced to the cutting edge in the visual recognition task. Transf...
Vision transformers (ViTs) have recently obtained success in many applications, but their intensive ...
Vision Transformers (ViT) have been shown to attain highly competitive performance for a wide range ...
The great success of transformer-based models in natural language processing (NLP) has led to variou...
Network quantization significantly reduces model inference complexity and has been widely used in re...
In this paper, we propose a fully differentiable quantization method for vision transformer (ViT) na...
Vision transformers have recently gained great success on various computer vision tasks; nevertheles...
Data-free quantization can potentially address data privacy and security concerns in model compressi...
Vision Transformer (ViT) architectures are becoming increasingly popular and widely employed to tack...
Attention-based neural networks such as the Vision Transformer (ViT) have recently attained state-of...
Vision Transformers (ViTs) have achieved state-of-the-art performance on various computer vision app...
Vision transformers (ViTs) have recently obtained success in many applications, but their intensive ...
While neural networks have been remarkably successful in a wide array of applications, implementing ...
The neural network quantization is highly desired procedure to perform before running neural network...
We attempt to reduce the computational costs in vision transformers (ViTs), which increase quadratic...
The vision transformer (ViT) has advanced to the cutting edge in the visual recognition task. Transf...
Vision transformers (ViTs) have recently obtained success in many applications, but their intensive ...
Vision Transformers (ViT) have been shown to attain highly competitive performance for a wide range ...
The great success of transformer-based models in natural language processing (NLP) has led to variou...