Skip to content

Commit 4df7e2d

Browse files
committed
update
1 parent cba5875 commit 4df7e2d

File tree

1 file changed

+4
-0
lines changed

1 file changed

+4
-0
lines changed

_posts/2025-12-18-vllm-omni-diffusion-cache-acceleration.md renamed to _posts/2025-12-19-vllm-omni-diffusion-cache-acceleration.md

Lines changed: 4 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -88,6 +88,8 @@ For image editing tasks, Cache-DiT shines even brighter. On **Qwen-Image-Edit**,
8888
</div>
8989
</div>
9090

91+
These caching optimization techniques show equally impressive results on heterogeneous platforms like Ascend NPU. For instance, Qwen-Image-Edit inference on Ascend NPU was accelerated using Cache-DiT from 142.38s down to 64.07s, achieving over a 2.2x speedup.
92+
9193
## Supported Models
9294

9395
| Model | TeaCache | Cache-DiT |
@@ -138,3 +140,5 @@ Ready to speed up your diffusion pipelines? Check out our detailed documentation
138140

139141
* [Cache-DiT Acceleration Guide](https://docs.vllm.ai/projects/vllm-omni/en/latest/user_guide/acceleration/cache_dit_acceleration/)
140142
* [TeaCache Guide](https://docs.vllm.ai/projects/vllm-omni/en/latest/user_guide/acceleration/teacache/)
143+
144+
Beyond caching, we are also actively developing optimizations in parallelization, kernel fusion, and quantization. Stay tuned for more powerful features!

0 commit comments

Comments
 (0)