HyperAI

Efficient Vits On Imagenet 1K With Deit S

Métriques

GFLOPs
Top 1 Accuracy

Résultats

Résultats de performance de divers modèles sur ce benchmark

Nom du modèle
GFLOPs
Top 1 Accuracy
Paper TitleRepository
HVT-S-12.778.3Scalable Vision Transformers with Hierarchical Pooling
DynamicViT (70%)2.979.3DynamicViT: Efficient Vision Transformers with Dynamic Token Sparsification
LTMP (45%)2.378.6Learned Thresholds Token Merging and Pruning for Vision Transformers
EViT (80%)3.579.8Not All Patches are What You Need: Expediting Vision Transformers via Token Reorganizations
BAT (70%)3.079.6Beyond Attentive Tokens: Incorporating Token Importance and Diversity for Efficient Vision Transformers
MCTF ($r=18$)2.479.9Multi-criteria Token Fusion with One-step-ahead Attention for Efficient Vision Transformers
BAT (60%)2.679.3Beyond Attentive Tokens: Incorporating Token Importance and Diversity for Efficient Vision Transformers
IA-RED$^2$3.279.1IA-RED$^2$: Interpretability-Aware Redundancy Reduction for Vision Transformers-
AS-DeiT-S (50%)2.378.7Adaptive Sparse ViT: Towards Learnable Adaptive Token Pruning by Fully Exploiting Self-Attention-
S$^2$ViTE3.279.2Chasing Sparsity in Vision Transformers: An End-to-End Exploration
BAT (20%)1.676.4Beyond Attentive Tokens: Incorporating Token Importance and Diversity for Efficient Vision Transformers
DPS-ViT2.479.5Patch Slimming for Efficient Vision Transformers-
LTMP (80%)3.879.8Learned Thresholds Token Merging and Pruning for Vision Transformers
SPViT3.378.3Pruning Self-attentions into Convolutional Layers in Single Path
DiffRate2.979.8DiffRate : Differentiable Compression Rate for Efficient Vision Transformers
AS-DeiT-S (65%)3.079.6Adaptive Sparse ViT: Towards Learnable Adaptive Token Pruning by Fully Exploiting Self-Attention-
Base (DeiT-S)4.679.8Training data-efficient image transformers & distillation through attention-
EViT (90%)4.079.8Not All Patches are What You Need: Expediting Vision Transformers via Token Reorganizations
EViT (60%)2.678.9Not All Patches are What You Need: Expediting Vision Transformers via Token Reorganizations
PPT2.979.8PPT: Token Pruning and Pooling for Efficient Vision Transformers
0 of 41 row(s) selected.