|
1 | | -# 📒Awesome-SD-Distributed-Inference |
| 1 | +# 📒Awesome-SD-Distributed/Caching-Inference |
2 | 2 | <div align='left'> |
3 | 3 | <img src=https://cdn.rawgit.com/sindresorhus/awesome/d7305f38d29fed78fa85652e3a63e154dd8e8829/media/badge.svg > |
4 | 4 | <img src=https://img.shields.io/github/forks/DefTruth/Awesome-SD-Distributed-Inference.svg?style=social > |
|
8 | 8 | <img src=https://img.shields.io/badge/License-GPLv3.0-turquoise.svg > |
9 | 9 | </div> |
10 | 10 |
|
11 | | -📒A **small** curated list of Awesome **SD/DiT/ViT/Diffusion** **Distributed Inference(Multi-GPUs)** Paper with codes. For Awesome LLM Inference, please check 📖[Awesome-LLM-Inference](https://github.com/DefTruth/Awesome-LLM-Inference)  |
| 11 | +📒A **small** curated list of Awesome **SD/DiT/ViT/Diffusion** **Distributed/Caching Inference** Paper with codes. For Awesome LLM Inference, please check 📖[Awesome-LLM-Inference](https://github.com/DefTruth/Awesome-LLM-Inference)  |
12 | 12 |
|
13 | 13 |
|
14 | 14 | ## ©️Citations |
15 | 15 |
|
16 | 16 | ```BibTeX |
17 | | -@misc{Awesome-SD-Distributed-Inference@2024, |
18 | | - title={Awesome-SD-Distributed-Inference: A small curated list of Awesome SD/DiT/ViT/Diffusion Distributed Inference Paper with codes}, |
19 | | - url={https://github.com/DefTruth/Awesome-SD-Distributed-Inference}, |
| 17 | +@misc{Awesome-SD-Distributed/Caching-Inference@2024, |
| 18 | + title={Awesome-SD-Distributed/Caching-Inference: A small curated list of Awesome SD/DiT/ViT/Diffusion Distributed/Caching Inference Paper with codes}, |
| 19 | + url={https://github.com/DefTruth/Awesome-SD-Distributed/Caching-Inference}, |
20 | 20 | note={Open-source software available at https://github.com/DefTruth/Awesome-SD-Distributed-Inference}, |
21 | 21 | author={DefTruth}, |
22 | 22 | year={2024} |
23 | 23 | } |
24 | 24 | ``` |
25 | 25 |
|
26 | | -## 📙Awesome SD Distributed Inference Papers with Codes |
| 26 | +## 📙Awesome SD Distributed/Caching Inference Papers with Codes |
| 27 | + |
| 28 | +- **UNet Based (DeepCache)** |
| 29 | + |
| 30 | +<img width="1645" alt="image" src="https://github.com/user-attachments/assets/a7257462-80d3-40af-a4ce-3550508fabe7"> |
| 31 | + |
| 32 | + |
| 33 | +- **DiT Based (Fast-Forward Caching)** |
| 34 | +<img width="1119" alt="image" src="https://github.com/user-attachments/assets/fad8f187-d4ac-4290-9943-7b34116fed05"> |
| 35 | + |
27 | 36 |
|
28 | 37 | |Date|Title|Paper|Code|Recom| |
29 | 38 | |:---:|:---:|:---:|:---:|:---:| |
| 39 | +|2023.05|🔥🔥[**Cache-Enabled Sparse Diffusion**] Accelerating Text-to-Image Editing via Cache-Enabled Sparse Diffusion Inference(@pku.edu.cn etc)|[[pdf]](https://arxiv.org/pdf/2305.17423) |⚠️|⭐️⭐️ | |
30 | 40 | |2023.12|🔥🔥[**DeepCache**] DeepCache: Accelerating Diffusion Models for Free(@nus.edu)|[[pdf]](https://arxiv.org/pdf/2312.00858) | [[DeepCache]](https://github.com/horseee/DeepCache) | ⭐️⭐️ | |
31 | 41 | |2023.12|🔥🔥[**Block Caching**] Cache Me if You Can: Accelerating Diffusion Models through Block Caching(@Meta GenAI etc)|[[pdf]](https://arxiv.org/pdf/2312.03209) |⚠️|⭐️⭐️ | |
| 42 | +|2023.12|🔥🔥[**Approximate Caching**] Approximate Caching for Efficiently Serving Diffusion Models(@Adobe)|[[pdf]](https://arxiv.org/pdf/2312.04429) |⚠️|⭐️⭐️ | |
32 | 43 | |2024.02|🔥🔥[**DistriFusion**] DistriFusion: Distributed Parallel Inference for High-Resolution Diffusion Models(@MIT etc)|[[pdf]](https://arxiv.org/abs/2402.19481) | [[distrifuser]](https://github.com/mit-han-lab/distrifuser) | ⭐️⭐️ | |
33 | 44 | |2024.05|🔥🔥[**PipeFusion**] PipeFusion: Displaced Patch Pipeline Parallelism for Inference of Diffusion Transformer Models(@Tencent etc)|[[pdf]](https://arxiv.org/pdf/2405.14430) | [[PipeFusion]](https://github.com/PipeFusion/PipeFusion) | ⭐️⭐️ | |
34 | 45 | |2024.06| 🔥🔥[**AsyncDiff**] AsyncDiff: Parallelizing Diffusion Models by Asynchronous Denoising(@nus.edu) | [[pdf]](https://arxiv.org/pdf/2406.06911) | [[AsyncDiff]](https://github.com/czg1225/AsyncDiff) | ⭐️⭐️ | |
35 | 46 | |2024.06| 🔥🔥[**Layer Caching**] Learning-to-Cache: Accelerating Diffusion Transformer via Layer Caching(@nus.edu) | [[pdf]](https://arxiv.org/pdf/2406.01733) | [[learning-to-cache]](https://github.com/horseee/learning-to-cache/) | ⭐️⭐️ | |
36 | 47 | |2024.05 | 🔥🔥[**TensorRT-LLM SDXL**] SDXL Distributed Inference with TensorRT-LLM and synchronous comm(@Zars19) | [[pdf]](https://arxiv.org/abs/2402.19481) | [[SDXL-TensorRT-LLM]](https://github.com/NVIDIA/TensorRT-LLM/pull/1514) | ⭐️⭐️ | |
37 | 48 | |2024.06| 🔥🔥[**Clip Parallelism**] Video-Infinity: Distributed Long Video Generation(@nus.edu)|[[pdf]](https://arxiv.org/pdf/2406.16260) | [[Video-Infinity]](https://github.com/Yuanshi9815/Video-Infinity) |⭐️⭐️ | |
38 | 49 | |2024.05| 🔥🔥[**FIFO-Diffusion**] FIFO-Diffusion: Generating Infinite Videos from Text without Training(@Seoul National University)|[[pdf]](https://arxiv.org/pdf/2405.11473) | [[FIFO-Diffusion]](https://github.com/jjihwan/FIFO-Diffusion_public)  |⭐️⭐️ | |
39 | | -|2024.07|🔥 [**ElasticCache-LVLM**] Efficient Inference of Vision Instruction-Following Models with Elastic Cache(@Tsinghua University etc)|[[pdf]](https://arxiv.org/pdf/2407.18121)|[[ElasticCache]](https://github.com/liuzuyan/ElasticCache) |⭐️ | |
| 50 | +|2024.07|🔥[**ElasticCache-LVLM**] Efficient Inference of Vision Instruction-Following Models with Elastic Cache(@Tsinghua University etc)|[[pdf]](https://arxiv.org/pdf/2407.18121)|[[ElasticCache]](https://github.com/liuzuyan/ElasticCache) |⭐️ | |
| 51 | +|2024.07| 🔥🔥[**Fast-Forward Caching(DiT)**] FORA: Fast-Forward Caching in Diffusion Transformer Acceleration(@microsoft.com etc) | [[pdf]](https://arxiv.org/pdf/2407.01425) | [[FORA]](https://github.com/prathebaselva/FORA) |⭐️⭐️ | |
40 | 52 |
|
41 | 53 | ## ©️License |
42 | 54 |
|
|
0 commit comments