view post Post 3221 Mistral's new Ministral 3 models can now be Run & Fine-tuned locally! (16GB RAM)Ministral 3 have vision support and the best-in-class performance for their sizes.14B Instruct GGUF: unsloth/Ministral-3-14B-Instruct-2512-GGUF14B Reasoning GGUF: unsloth/Ministral-3-14B-Reasoning-2512-GGUFπ± Step-by-step Guide: https://docs.unsloth.ai/new/ministral-3All GGUFs, BnB, FP8 etc. variants uploads: https://huggingface.co/collections/unsloth/ministral-3 See translation 3 replies Β· π₯ 17 17 π€ 6 6 β€οΈ 5 5 π 3 3 + Reply
view post Post 8226 Qwen3-Next can now be Run locally! (30GB RAM)Instruct GGUF: unsloth/Qwen3-Next-80B-A3B-Instruct-GGUFThe models come in Thinking and Instruct versions and utilize a new architecture, allowing it to have ~10x faster inference than Qwen32B.π Step-by-step Guide: https://docs.unsloth.ai/models/qwen3-nextThinking GGUF: unsloth/Qwen3-Next-80B-A3B-Thinking-GGUF See translation π₯ 37 37 β€οΈ 11 11 π 7 7 π€ 3 3 + Reply
view post Post 4203 You can now run Kimi K2 Thinking locally with our Dynamic 1-bit GGUFs: unsloth/Kimi-K2-Thinking-GGUFWe shrank the 1T model to 245GB (-62%) & retained ~85% of accuracy on Aider Polyglot. Run on >247GB RAM for fast inference.We also collaborated with the Moonshot AI Kimi team on a system prompt fix! π₯°Guide + fix details: https://docs.unsloth.ai/models/kimi-k2-thinking-how-to-run-locally See translation β€οΈ 10 10 π 9 9 π₯ 6 6 π€ 4 4 π€― 3 3 + Reply
view post Post 6470 Run DeepSeek-V3.1 locally on 170GB RAM with Dynamic 1-bit GGUFs!πGGUFs: unsloth/DeepSeek-V3.1-GGUFThe 715GB model gets reduced to 170GB (-80% size) by smartly quantizing layers.The 1-bit GGUF passes all our code tests & we fixed the chat template for llama.cpp supported backends.Guide: https://docs.unsloth.ai/basics/deepseek-v3.1 See translation β€οΈ 19 19 π₯ 9 9 π 5 5 + Reply
view post Post 5537 Run OpenAI's new gpt-oss models locally with Unsloth GGUFs! π₯π¦₯20b GGUF: unsloth/gpt-oss-20b-GGUF120b GGUF: unsloth/gpt-oss-120b-GGUFModel will run on 14GB RAM for 20b and 66GB for 120b. See translation 2 replies Β· β€οΈ 21 21 π₯ 6 6 π 5 5 + Reply
view post Post 3665 It's Qwen3 week! π We uploaded Dynamic 2-bit GGUFs for:Qwen3-Coder: unsloth/Qwen3-Coder-480B-A35B-Instruct-GGUFQwen3-2507: unsloth/Qwen3-235B-A22B-Instruct-2507-GGUFSo you can run them both locally!Guides are in model cards. See translation 1 reply Β· π€ 5 5 β€οΈ 4 4 π₯ 3 3 + Reply
view post Post 3884 Made some 245GB (80% size reduction) 1.8bit quants for Kimi K2! unsloth/Kimi-K2-Instruct-GGUF See translation π₯ 10 10 π€― 2 2 + Reply
view post Post 3972 We fixed more issues! Use --jinja for all!* Fixed Nanonets OCR-s unsloth/Nanonets-OCR-s-GGUF* Fixed THUDM GLM-4 unsloth/GLM-4-32B-0414-GGUF* DeepSeek Chimera v2 is uploading! unsloth/DeepSeek-TNG-R1T2-Chimera-GGUF See translation β€οΈ 3 3 π 3 3 π€ 2 2 + Reply
view post Post 3181 Gemma 3n finetuning is now 1.5x faster and uses 50% less VRAM in Unsloth!Click "Use this model" and click "Google Colab"! unsloth/gemma-3n-E4B-it unsloth/gemma-3n-E2B-ithttps://colab.research.google.com/github/unslothai/notebooks/blob/main/nb/Gemma3N_(4B)-Conversational.ipynb See translation 2 replies Β· π₯ 3 3 β€οΈ 3 3 π 2 2 + Reply
view post Post 1319 We updated lots of our GGUFs and uploaded many new ones!* unsloth/dots.llm1.inst-GGUF* unsloth/Jan-nano-GGUF* unsloth/Nanonets-OCR-s-GGUF* Updated and fixed Q8_0 upload for unsloth/DeepSeek-R1-0528-Qwen3-8B-GGUF* Added Q2_K_XL for unsloth/DeepSeek-R1-0528-GGUF* Updated and fixed Vision support for unsloth/Llama-4-Scout-17B-16E-Instruct-GGUF See translation π€ 2 2 π₯ 1 1 + Reply
view post Post 2524 Mistral releases Magistral, their new reasoning models! π₯GGUFs to run: unsloth/Magistral-Small-2506-GGUFMagistral-Small-2506 excels at mathematics and coding.You can run the 24B model locally with just 32GB RAM by using our Dynamic GGUFs. See translation π₯ 7 7 π€ 2 2 π 1 1 + Reply
view post Post 3857 New DeepSeek-R1-0528 1.65-bit Dynamic GGUF!Run the model locally even easier! Will fit on a 192GB Macbook and run at 7 tokens/s.DeepSeek-R1-0528 GGUFs: unsloth/DeepSeek-R1-0528-GGUFQwen3-8B DeepSeek-R1-0528 GGUFs: unsloth/DeepSeek-R1-0528-Qwen3-8B-GGUFAnd read our Guide: https://docs.unsloth.ai/basics/deepseek-r1-0528 See translation π₯ 10 10 β€οΈ 8 8 π 3 3 π€ 3 3 π€ 2 2 π€― 1 1 + Reply
view post Post 2331 π Qwen3 128K Context Length: We've released Dynamic 2.0 GGUFs + 4-bit safetensors!Fixed: Now works on any inference engine and fixed issues with the chat template.Qwen3 GGUFs:30B-A3B: unsloth/Qwen3-30B-A3B-GGUF235-A22B: unsloth/Qwen3-235B-A22B-GGUF32B: unsloth/Qwen3-32B-GGUFRead our guide on running Qwen3 here: https://docs.unsloth.ai/basics/qwen3-how-to-run-and-finetune128K Context Length:30B-A3B: unsloth/Qwen3-30B-A3B-128K-GGUF235-A22B: unsloth/Qwen3-235B-A22B-128K-GGUF32B: unsloth/Qwen3-32B-128K-GGUFAll Qwen3 uploads: unsloth/qwen3-680edabfb790c8c34a242f95 See translation β€οΈ 7 7 π€ 1 1 π 1 1 π₯ 1 1 + Reply