4-bit OmniQuant quantized version of FuseChat-Gemma-2-9B-Instruct for inference with the Private LLM app.
Base model