GLM-4.6V-exl3 / README.md
turboderp's picture
Update README.md
e06c0fd verified
---
license: mit
base_model: zai-org/GLM-4.6V
base_model_relation: quantized
quantized_by: turboderp
tags:
- exl3
---
EXL3 quants of [GLM-4.6V](https://huggingface.co/zai-org/GLM-4.6V)
⚠️ Requires ExLlamaV3 v0.0.18 (or v0.0.17 `dev` branch)
Base bitrates:
[2.00 bits per weight](https://huggingface.co/turboderp/GLM-4.6V-exl3/tree/2.00bpw)
[3.00 bits per weight](https://huggingface.co/turboderp/GLM-4.6V-exl3/tree/3.00bpw)
[4.00 bits per weight](https://huggingface.co/turboderp/GLM-4.6V-exl3/tree/4.00bpw)
[5.00 bits per weight](https://huggingface.co/turboderp/GLM-4.6V-exl3/tree/5.00bpw)
Optimized:
[2.13 bits per weight](https://huggingface.co/turboderp/GLM-4.6V-exl3/tree/2.13bpw)
[2.32 bits per weight](https://huggingface.co/turboderp/GLM-4.6V-exl3/tree/2.32bpw)
[2.55 bits per weight](https://huggingface.co/turboderp/GLM-4.6V-exl3/tree/2.55bpw)
[2.80 bits per weight](https://huggingface.co/turboderp/GLM-4.6V-exl3/tree/2.80bpw)
[3.13 bits per weight](https://huggingface.co/turboderp/GLM-4.6V-exl3/tree/3.13bpw)
[3.55 bits per weight](https://huggingface.co/turboderp/GLM-4.6V-exl3/tree/3.55bpw)
[4.07 bits per weight](https://huggingface.co/turboderp/GLM-4.6V-exl3/tree/4.07bpw)