v0.5.3
What's Changed
- Add Quantization topic by @DefTruth in #26
- 🔥[DiTFastAttnV2] Head-wise Attention Compression for Multi-Modality Diffusion Transformers by @DefTruth in #27
- add quantization paper mixdq & viditq by @A-suozhang in #28
- 🔥🔥[Increment-Calibrated Cache] Accelerating Diffusion Transformer via Increment-Calibrated Caching with Channel-Aware Singular Value Decomposition by @DefTruth in #29
- SageAttention3: Microscaling FP4 Attention for Inference and An Exploration of 8-bit Training by @DefTruth in #30
- FastCache: Fast Caching for Diffusion Transformer Through Learnable Linear Approximation by @DefTruth in #31
New Contributors
- @A-suozhang made their first contribution in #28
Full Changelog: v0.5.2...v0.5.3