FP8-Dynamic quantization using llmcompressor. Run with:
vllm serve leon-se/gemma-3-27b-it-FP8-Dynamic --max-model-len 4096
- Downloads last month
 - 2,456
 
	Inference Providers
	NEW
	
	
	This model isn't deployed by any Inference Provider.
	🙋
			1
		Ask for provider support