Quant-dLLM: 2‑Bit Post‑Training Compression for Diffusion LLMs
Quant-dLLM introduces a framework for 2‑bit post‑training quantization of diffusion large language models, preserving performance. The code and pretrained models will be open‑sourced on GitHub. getnews.me/quant-dllm-2-bit-post-tr... #quantdllm #diffusionllm
0
0
0
0