Apertus-8B-Instruct-2509-FP8-Dynamic
This is an FP8 dynamically quantized version of swiss-ai/Apertus-8B-Instruct-2509 using llm-compressor.
Quantization Details
- Quantization Scheme: FP8_dynamic
- Method: Dynamic quantization of weights and activations to FP8 format
- Targets: All Linear layers
- Ignored Layers:
lm_head(kept in higher precision for better output quality) - Tool: llm-compressor (Neural Magic)
- Downloads last month
- 31
Model tree for starbix/Apertus-8B-Instruct-2509-FP8_dynamic
Base model
swiss-ai/Apertus-8B-2509
Finetuned
swiss-ai/Apertus-8B-Instruct-2509