This LLM Compressor v0.7.0 release introduces major new features and enhancements:
- Apply multiple compressors to a single model for mixed-precision quantization
- Support for DeepSeekV3-style block FP8 quantization
- Configurable Observer arguments in
config.yaml
- Expanded Mixture of Experts (MoE) calibration support, including NVFP4
- Llama4 quantization support with vLLM compatibility
- Simplified and unified Recipe classes for easier usage and debugging