Papers
arxiv:2512.16093

TurboDiffusion: Accelerating Video Diffusion Models by 100-200 Times

Published on Dec 18
· Submitted by
Jintao Zhang
on Dec 25
#2 Paper of the day
Authors:
,
,
,
,
,
,
,

Abstract

TurboDiffusion accelerates video generation by 100-200x using attention acceleration, step distillation, and quantization, while maintaining video quality.

AI-generated summary

We introduce TurboDiffusion, a video generation acceleration framework that can speed up end-to-end diffusion generation by 100-200x while maintaining video quality. TurboDiffusion mainly relies on several components for acceleration: (1) Attention acceleration: TurboDiffusion uses low-bit SageAttention and trainable Sparse-Linear Attention (SLA) to speed up attention computation. (2) Step distillation: TurboDiffusion adopts rCM for efficient step distillation. (3) W8A8 quantization: TurboDiffusion quantizes model parameters and activations to 8 bits to accelerate linear layers and compress the model. In addition, TurboDiffusion incorporates several other engineering optimizations. We conduct experiments on the Wan2.2-I2V-14B-720P, Wan2.1-T2V-1.3B-480P, Wan2.1-T2V-14B-720P, and Wan2.1-T2V-14B-480P models. Experimental results show that TurboDiffusion achieves 100-200x speedup for video generation even on a single RTX 5090 GPU, while maintaining comparable video quality. The GitHub repository, which includes model checkpoints and easy-to-use code, is available at https://github.com/thu-ml/TurboDiffusion.

Community

TurboDiffusion: 100–200× acceleration in video generation on a single RTX 5090.

A high-quality 5-second video can be generated in just 1.9 seconds. Efficient inference code, as well as model parameters (checkpoints) for TurboWan2.2/2.1 for Text-to-Video and Image-to-Video generation, have been open-sourced for one-click generation.

The core techniques are: SageAttention + Sparse-Linear Attention (SLA) + rCM + W8A8.

Github: https://github.com/thu-ml/TurboDiffusion

Technical Report: https://jt-zhang.github.io/files/TurboDiffusion_Technical_Report.pdf


2
3
4

Sign up or log in to comment

Models citing this paper 8

Browse 8 models citing this paper

Datasets citing this paper 0

No dataset linking this paper

Cite arxiv.org/abs/2512.16093 in a dataset README.md to link it from this page.

Spaces citing this paper 0

No Space linking this paper

Cite arxiv.org/abs/2512.16093 in a Space README.md to link it from this page.

Collections including this paper 2