Official code for paper: [CLS] Attention is All You Need for Training-Free Visual Token Pruning: Make VLM Inference Faster.
-
Updated
Dec 14, 2024 - Python
Official code for paper: [CLS] Attention is All You Need for Training-Free Visual Token Pruning: Make VLM Inference Faster.
ICML2025: AsymRnR: Video Diffusion Transformers Acceleration with Asymmetric Reduction and Restoration
Add a description, image, and links to the training-free-acceleration topic page so that developers can more easily learn about it.
To associate your repository with the training-free-acceleration topic, visit your repo's landing page and select "manage topics."