Skip to content

v0.5.3

Choose a tag to compare

@DefTruth DefTruth released this 03 Jun 06:41
· 30 commits to main since this release
c6a12c8

What's Changed

  • Add Quantization topic by @DefTruth in #26
  • 🔥[DiTFastAttnV2] Head-wise Attention Compression for Multi-Modality Diffusion Transformers by @DefTruth in #27
  • add quantization paper mixdq & viditq by @A-suozhang in #28
  • 🔥🔥[Increment-Calibrated Cache] Accelerating Diffusion Transformer via Increment-Calibrated Caching with Channel-Aware Singular Value Decomposition by @DefTruth in #29
  • SageAttention3: Microscaling FP4 Attention for Inference and An Exploration of 8-bit Training by @DefTruth in #30
  • FastCache: Fast Caching for Diffusion Transformer Through Learnable Linear Approximation by @DefTruth in #31

New Contributors

Full Changelog: v0.5.2...v0.5.3