Qwen3-VL Collection Qwen's new multimodal vision models in GGUF, safetensor, and dynamic Unsloth formats. • 56 items • Updated 1 day ago • 8
view post Post 6008 Run DeepSeek-V3.1 locally on 170GB RAM with Dynamic 1-bit GGUFs!🐋GGUFs: unsloth/DeepSeek-V3.1-GGUFThe 715GB model gets reduced to 170GB (-80% size) by smartly quantizing layers.The 1-bit GGUF passes all our code tests & we fixed the chat template for llama.cpp supported backends.Guide: https://docs.unsloth.ai/basics/deepseek-v3.1 See translation ❤️ 18 18 🔥 9 9 🚀 5 5 + Reply
view post Post 5207 Run OpenAI's new gpt-oss models locally with Unsloth GGUFs! 🔥🦥20b GGUF: unsloth/gpt-oss-20b-GGUF120b GGUF: unsloth/gpt-oss-120b-GGUFModel will run on 14GB RAM for 20b and 66GB for 120b. See translation 2 replies · ❤️ 20 20 🔥 6 6 🚀 5 5 + Reply
view post Post 3403 It's Qwen3 week! 💜 We uploaded Dynamic 2-bit GGUFs for:Qwen3-Coder: unsloth/Qwen3-Coder-480B-A35B-Instruct-GGUFQwen3-2507: unsloth/Qwen3-235B-A22B-Instruct-2507-GGUFSo you can run them both locally!Guides are in model cards. See translation 1 reply · 🤗 5 5 ❤️ 4 4 🔥 3 3 + Reply
view post Post 3652 Made some 245GB (80% size reduction) 1.8bit quants for Kimi K2! unsloth/Kimi-K2-Instruct-GGUF See translation 🔥 10 10 🤯 2 2 + Reply
view post Post 3789 We fixed more issues! Use --jinja for all!* Fixed Nanonets OCR-s unsloth/Nanonets-OCR-s-GGUF* Fixed THUDM GLM-4 unsloth/GLM-4-32B-0414-GGUF* DeepSeek Chimera v2 is uploading! unsloth/DeepSeek-TNG-R1T2-Chimera-GGUF See translation ❤️ 3 3 👍 3 3 🤗 2 2 + Reply
view post Post 3036 Gemma 3n finetuning is now 1.5x faster and uses 50% less VRAM in Unsloth!Click "Use this model" and click "Google Colab"! unsloth/gemma-3n-E4B-it unsloth/gemma-3n-E2B-ithttps://colab.research.google.com/github/unslothai/notebooks/blob/main/nb/Gemma3N_(4B)-Conversational.ipynb See translation 2 replies · 🔥 3 3 ❤️ 3 3 🚀 2 2 + Reply