'Make knowledge free for everyone'

Original INT4 model has been dequantized with my own custom script:

DQ_int4-to-bf16_dequant (inspired by the deepseek V3 dequant script)

Test/Proof

kimi-think-proof

Zero Short Hexa-ball test, generated code by the Q3 quant produced:

Kimi-Think_Hexa-Ball_test

Quantized version of: moonshotai/Kimi-K2-Thinking Buy Me a Coffee at ko-fi.com

Downloads last month
1,771
GGUF
Model size
1T params
Architecture
deepseek2
Hardware compatibility
Log In to view the estimation

2-bit

3-bit

4-bit

Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for DevQuasar/moonshotai.Kimi-K2-Thinking-GGUF

Quantized
(5)
this model

Collection including DevQuasar/moonshotai.Kimi-K2-Thinking-GGUF