AI & ML interests

Open Source AI πŸ’š

Recent Activity

danielhanchenΒ  updated a model about 6 hours ago
unsloth/GLM-4.6V-Flash-GGUF
danielhanchenΒ  updated a model about 8 hours ago
unsloth/GLM-4.6V-Flash
danielhanchenΒ  published a model about 8 hours ago
unsloth/GLM-4.6V-Flash
View all activity

danielhanchenΒ 
posted an update 6 days ago
view post
Post
3221
Mistral's new Ministral 3 models can now be Run & Fine-tuned locally! (16GB RAM)
Ministral 3 have vision support and the best-in-class performance for their sizes.
14B Instruct GGUF: unsloth/Ministral-3-14B-Instruct-2512-GGUF
14B Reasoning GGUF: unsloth/Ministral-3-14B-Reasoning-2512-GGUF

🐱 Step-by-step Guide: https://docs.unsloth.ai/new/ministral-3
All GGUFs, BnB, FP8 etc. variants uploads: https://huggingface.co/collections/unsloth/ministral-3
Β·
danielhanchenΒ 
posted an update 10 days ago
danielhanchenΒ 
posted an update about 1 month ago
view post
Post
4203
You can now run Kimi K2 Thinking locally with our Dynamic 1-bit GGUFs: unsloth/Kimi-K2-Thinking-GGUF

We shrank the 1T model to 245GB (-62%) & retained ~85% of accuracy on Aider Polyglot. Run on >247GB RAM for fast inference.

We also collaborated with the Moonshot AI Kimi team on a system prompt fix! πŸ₯°

Guide + fix details: https://docs.unsloth.ai/models/kimi-k2-thinking-how-to-run-locally
danielhanchenΒ 
posted an update 4 months ago
view post
Post
6470
Run DeepSeek-V3.1 locally on 170GB RAM with Dynamic 1-bit GGUFs!πŸ‹
GGUFs: unsloth/DeepSeek-V3.1-GGUF

The 715GB model gets reduced to 170GB (-80% size) by smartly quantizing layers.

The 1-bit GGUF passes all our code tests & we fixed the chat template for llama.cpp supported backends.

Guide: https://docs.unsloth.ai/basics/deepseek-v3.1
danielhanchenΒ 
posted an update 4 months ago
danielhanchenΒ 
posted an update 5 months ago
danielhanchenΒ 
posted an update 5 months ago
danielhanchenΒ 
posted an update 5 months ago
danielhanchenΒ 
posted an update 5 months ago
danielhanchenΒ 
posted an update 6 months ago
danielhanchenΒ 
posted an update 6 months ago
view post
Post
2524
Mistral releases Magistral, their new reasoning models! πŸ”₯
GGUFs to run: unsloth/Magistral-Small-2506-GGUF

Magistral-Small-2506 excels at mathematics and coding.

You can run the 24B model locally with just 32GB RAM by using our Dynamic GGUFs.
danielhanchenΒ 
posted an update 6 months ago
danielhanchenΒ 
posted an update 7 months ago
view post
Post
2331
πŸ’œ Qwen3 128K Context Length: We've released Dynamic 2.0 GGUFs + 4-bit safetensors!
Fixed: Now works on any inference engine and fixed issues with the chat template.
Qwen3 GGUFs:
30B-A3B: unsloth/Qwen3-30B-A3B-GGUF
235-A22B: unsloth/Qwen3-235B-A22B-GGUF
32B: unsloth/Qwen3-32B-GGUF

Read our guide on running Qwen3 here: https://docs.unsloth.ai/basics/qwen3-how-to-run-and-finetune

128K Context Length:
30B-A3B: unsloth/Qwen3-30B-A3B-128K-GGUF
235-A22B: unsloth/Qwen3-235B-A22B-128K-GGUF
32B: unsloth/Qwen3-32B-128K-GGUF

All Qwen3 uploads: unsloth/qwen3-680edabfb790c8c34a242f95