|
27 | 27 |
|
28 | 28 | |Date|Title|Paper|Code|Recom| |
29 | 29 | |:---:|:---:|:---:|:---:|:---:| |
30 | | -|2023.12|🔥🔥[DeepCache] DeepCache: Accelerating Diffusion Models for Free(@nus.edu)|[[pdf]](https://arxiv.org/pdf/2312.00858) | [[DeepCache]](https://github.com/horseee/DeepCache) | ⭐️⭐️ | |
31 | | -|2024.02|🔥🔥[DistriFusion] DistriFusion: Distributed Parallel Inference for High-Resolution Diffusion Models(@MIT etc)|[[pdf]](https://arxiv.org/abs/2402.19481) | [[distrifuser]](https://github.com/mit-han-lab/distrifuser) | ⭐️⭐️ | |
32 | | -|2024.05|🔥🔥[PipeFusion] PipeFusion: Displaced Patch Pipeline Parallelism for Inference of Diffusion Transformer Models(@Tencent etc)|[[pdf]](https://arxiv.org/pdf/2405.14430) | [[PipeFusion]](https://github.com/PipeFusion/PipeFusion) | ⭐️⭐️ | |
33 | | -|2024.06| 🔥🔥[AsyncDiff] AsyncDiff: Parallelizing Diffusion Models by Asynchronous Denoising(@nus.edu) | [[pdf]](https://arxiv.org/pdf/2406.06911) | [[AsyncDiff]](https://github.com/czg1225/AsyncDiff) | ⭐️⭐️ | |
34 | | -|2024.06| 🔥🔥[Layer Caching] Learning-to-Cache: Accelerating Diffusion Transformer via Layer Caching(@nus.edu) | [[pdf]](https://arxiv.org/pdf/2406.01733) | [[learning-to-cache]](https://github.com/horseee/learning-to-cache/) | ⭐️⭐️ | |
35 | | -|2024.05 | 🔥🔥[TensorRT-LLM SDXL] SDXL Distributed Inference in TensorRT-LLM with synchronous comm(@Zars19) | [[pdf]](https://arxiv.org/abs/2402.19481) | [[SDXL-TensorRT-LLM]](https://github.com/NVIDIA/TensorRT-LLM/pull/1514) | ⭐️⭐️ | |
| 30 | +|2023.12|🔥🔥[**DeepCache**] DeepCache: Accelerating Diffusion Models for Free(@nus.edu)|[[pdf]](https://arxiv.org/pdf/2312.00858) | [[DeepCache]](https://github.com/horseee/DeepCache) | ⭐️⭐️ | |
| 31 | +|2024.02|🔥🔥[**DistriFusion**] DistriFusion: Distributed Parallel Inference for High-Resolution Diffusion Models(@MIT etc)|[[pdf]](https://arxiv.org/abs/2402.19481) | [[distrifuser]](https://github.com/mit-han-lab/distrifuser) | ⭐️⭐️ | |
| 32 | +|2024.05|🔥🔥[**PipeFusion**] PipeFusion: Displaced Patch Pipeline Parallelism for Inference of Diffusion Transformer Models(@Tencent etc)|[[pdf]](https://arxiv.org/pdf/2405.14430) | [[PipeFusion]](https://github.com/PipeFusion/PipeFusion) | ⭐️⭐️ | |
| 33 | +|2024.06| 🔥🔥[**AsyncDiff**] AsyncDiff: Parallelizing Diffusion Models by Asynchronous Denoising(@nus.edu) | [[pdf]](https://arxiv.org/pdf/2406.06911) | [[AsyncDiff]](https://github.com/czg1225/AsyncDiff) | ⭐️⭐️ | |
| 34 | +|2024.06| 🔥🔥[**Layer Caching**] Learning-to-Cache: Accelerating Diffusion Transformer via Layer Caching(@nus.edu) | [[pdf]](https://arxiv.org/pdf/2406.01733) | [[learning-to-cache]](https://github.com/horseee/learning-to-cache/) | ⭐️⭐️ | |
| 35 | +|2024.05 | 🔥🔥[**TensorRT-LLM SDXL**] SDXL Distributed Inference in TensorRT-LLM with synchronous comm(@Zars19) | [[pdf]](https://arxiv.org/abs/2402.19481) | [[SDXL-TensorRT-LLM]](https://github.com/NVIDIA/TensorRT-LLM/pull/1514) | ⭐️⭐️ | |
36 | 36 |
|
37 | 37 |
|
38 | 38 | ## ©️License |
|
0 commit comments