This is a Q8_0 GGUF quantization of inclusionAI/Ling-flash-2.0.

Downloads last month
37
GGUF
Model size
103B params
Architecture
bailingmoe2
Hardware compatibility
Log In to view the estimation

8-bit

Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support

Model tree for ddh0/Ling-flash-2.0-Q8_0.gguf

Quantized
(15)
this model