AI & ML interests

Open Source AI πŸ’š

Recent Activity

danielhanchenΒ 
posted an update about 4 hours ago
view post
Post
133
You can now run Kimi K2 Thinking locally with our Dynamic 1-bit GGUFs: unsloth/Kimi-K2-Thinking-GGUF

We shrank the 1T model to 245GB (-62%) & retained ~85% of accuracy on Aider Polyglot. Run on >247GB RAM for fast inference.

We also collaborated with the Moonshot AI Kimi team on a system prompt fix! πŸ₯°

Guide + fix details: https://docs.unsloth.ai/models/kimi-k2-thinking-how-to-run-locally
danielhanchenΒ 
in unsloth/MiniMax-M2-GGUF about 9 hours ago

Pls MXFP4

πŸ”₯ 1
5
#4 opened 5 days ago by
Kirara702

No thinking tags when it runs?

9
#1 opened about 20 hours ago by
Disdrix
danielhanchenΒ 
posted an update 3 months ago
view post
Post
6155
Run DeepSeek-V3.1 locally on 170GB RAM with Dynamic 1-bit GGUFs!πŸ‹
GGUFs: unsloth/DeepSeek-V3.1-GGUF

The 715GB model gets reduced to 170GB (-80% size) by smartly quantizing layers.

The 1-bit GGUF passes all our code tests & we fixed the chat template for llama.cpp supported backends.

Guide: https://docs.unsloth.ai/basics/deepseek-v3.1
danielhanchenΒ 
posted an update 3 months ago
danielhanchenΒ 
posted an update 4 months ago
danielhanchenΒ 
posted an update 4 months ago
danielhanchenΒ 
posted an update 4 months ago
danielhanchenΒ 
posted an update 4 months ago
danielhanchenΒ 
posted an update 5 months ago
danielhanchenΒ 
posted an update 5 months ago
view post
Post
2446
Mistral releases Magistral, their new reasoning models! πŸ”₯
GGUFs to run: unsloth/Magistral-Small-2506-GGUF

Magistral-Small-2506 excels at mathematics and coding.

You can run the 24B model locally with just 32GB RAM by using our Dynamic GGUFs.
danielhanchenΒ 
posted an update 5 months ago
danielhanchenΒ 
posted an update 6 months ago
view post
Post
2287
πŸ’œ Qwen3 128K Context Length: We've released Dynamic 2.0 GGUFs + 4-bit safetensors!
Fixed: Now works on any inference engine and fixed issues with the chat template.
Qwen3 GGUFs:
30B-A3B: unsloth/Qwen3-30B-A3B-GGUF
235-A22B: unsloth/Qwen3-235B-A22B-GGUF
32B: unsloth/Qwen3-32B-GGUF

Read our guide on running Qwen3 here: https://docs.unsloth.ai/basics/qwen3-how-to-run-and-finetune

128K Context Length:
30B-A3B: unsloth/Qwen3-30B-A3B-128K-GGUF
235-A22B: unsloth/Qwen3-235B-A22B-128K-GGUF
32B: unsloth/Qwen3-32B-128K-GGUF

All Qwen3 uploads: unsloth/qwen3-680edabfb790c8c34a242f95