π§ Quem-v2-4b GGUFs
Quantized version of: rodrigomt/quem-V2-4b
π¦ Available GGUFs
| Format | Description |
|---|---|
| F16 | Full precision (16-bit), better quality, larger size βοΈ |
| Q8_K_XL | Quantized (8-bit XL variant, based on the quantization table of the unsloth model Qwen3-4B-Thinking-2507), smaller size, faster inference β‘ |
| Q4_K_XL | Quantized (4-bit XL variant, based on the quantization table of the unsloth model Qwen3-4B-Thinking-2507), smaller size, faster inference β‘ |
π Usage
Example with llama.cpp:
./main -m ./gguf-file-name.gguf -p "Hello world!"
- Downloads last month
- 14
Hardware compatibility
Log In
to view the estimation
4-bit
8-bit
16-bit
Model tree for rodrigomt/quem-V2-4b-GGUF
Base model
rodrigomt/quem-V2-4b