AI sources.twitter 5h ago 1 min read
PyTorch said on April 8 that MXFP8 and NVFP4 quantization with Diffusers and TorchAO can cut diffusion latency on NVIDIA B200 GPUs, with NVFP4 reaching up to 1.68x speedups. The accompanying blog frames selective quantization and regional compilation as the practical recipe for better latency-memory tradeoffs.