Very Large GGUFs
Collection
GGUF quantized versions of very large models - over 100B parameters
•
52 items
•
Updated
•
5
'Make knowledge free for everyone'
Original INT4 model has been dequantized with my own custom script:
DQ_int4-to-bf16_dequant (inspired by the deepseek V3 dequant script)
Quantized version of: moonshotai/Kimi-K2-Thinking
![]()
2-bit
3-bit
4-bit
Base model
moonshotai/Kimi-K2-Thinking