Advanced quantization toolkit for LLMs and VLMs. Native support for WOQ, MXFP4, NVFP4, GGUF, Adaptive Bits and seamless integration with Transformers, vLLM, SGLang, and TorchAO
-
Updated
Nov 21, 2025 - Python
Advanced quantization toolkit for LLMs and VLMs. Native support for WOQ, MXFP4, NVFP4, GGUF, Adaptive Bits and seamless integration with Transformers, vLLM, SGLang, and TorchAO
Add a description, image, and links to the mxfp4 topic page so that developers can more easily learn about it.
To associate your repository with the mxfp4 topic, visit your repo's landing page and select "manage topics."