Recent years have witnessed the great success of vision transformer (ViT), which has achieved state-of-the-art performance on multiple computer vision benchmarks. However, ViT models suffer from vast amounts of parameters and high computation cost, leading to difficult deployment on resource-constrained edge devices. Existing solutions mostly compress ViT models to a compact model but still cannot achieve real-time inference. To tackle this issue, we propose to explore the divisibility of transformer structure, and decompose the large ViT into multiple small models for collaborative inference at edge devices. Our objective is to achieve fast and energy-efficient collaborative inference while maintaining comparable accuracy compared with lar...
The past several years have witnessed the success of transformer-based models, and their scale and a...
In the past few years, transformers have achieved promising performances on various computer vision ...
Attention-based neural networks such as the Vision Transformer (ViT) have recently attained state-of...
Self-attention based models such as vision transformers (ViTs) have emerged as a very competitive ar...
With the success of Vision Transformers (ViTs) in computer vision tasks, recent arts try to optimize...
Vision Transformer (ViT) architectures are becoming increasingly popular and widely employed to tack...
Vision Transformers (ViTs) have shown impressive performance and have become a unified backbone for ...
Recently, Vision Transformer (ViT) has continuously established new milestones in the computer visio...
For time-critical IoT applications using deep learning, inference acceleration through distributed c...
Vision transformers (ViTs) have recently obtained success in many applications, but their intensive ...
This paper investigates task-oriented communication for multi-device cooperative edge inference, whe...
Vision Transformers (ViTs) have achieved state-of-the-art performance on various vision tasks. Howev...
Vision Transformers (ViTs) have achieved state-of-the-art performance on various computer vision app...
Cooperative inference in Mobile Edge Computing (MEC), achieved by deploying partitioned Deep Neural ...
We attempt to reduce the computational costs in vision transformers (ViTs), which increase quadratic...
The past several years have witnessed the success of transformer-based models, and their scale and a...
In the past few years, transformers have achieved promising performances on various computer vision ...
Attention-based neural networks such as the Vision Transformer (ViT) have recently attained state-of...
Self-attention based models such as vision transformers (ViTs) have emerged as a very competitive ar...
With the success of Vision Transformers (ViTs) in computer vision tasks, recent arts try to optimize...
Vision Transformer (ViT) architectures are becoming increasingly popular and widely employed to tack...
Vision Transformers (ViTs) have shown impressive performance and have become a unified backbone for ...
Recently, Vision Transformer (ViT) has continuously established new milestones in the computer visio...
For time-critical IoT applications using deep learning, inference acceleration through distributed c...
Vision transformers (ViTs) have recently obtained success in many applications, but their intensive ...
This paper investigates task-oriented communication for multi-device cooperative edge inference, whe...
Vision Transformers (ViTs) have achieved state-of-the-art performance on various vision tasks. Howev...
Vision Transformers (ViTs) have achieved state-of-the-art performance on various computer vision app...
Cooperative inference in Mobile Edge Computing (MEC), achieved by deploying partitioned Deep Neural ...
We attempt to reduce the computational costs in vision transformers (ViTs), which increase quadratic...
The past several years have witnessed the success of transformer-based models, and their scale and a...
In the past few years, transformers have achieved promising performances on various computer vision ...
Attention-based neural networks such as the Vision Transformer (ViT) have recently attained state-of...