These are quantizations of the model Gelato-30B-A3B
The imatrix has been used from mradermacher.
As most of the quants are available from the great mradermacher team, I will include here only the quants that are missing.
Usage Notes:
- Download the latest llama.cpp to use these quantizations.
- Try to use the best quality you can run.
- For the
mmprojfile, the F32 version is recommended for best results (F32 > BF16 > F16).
- Downloads last month
- 2,001
Hardware compatibility
Log In
to view the estimation
4-bit
Model tree for noctrex/Gelato-30B-A3B-i1-GGUF
Base model
Qwen/Qwen3-VL-30B-A3B-Instruct
Finetuned
mlfoundations/Gelato-30B-A3B