Quantized Qwen3-VL
Collection
4 items
•
Updated
This is Qwen/Qwen3-VL-8B-Instruct quantized with AutoRound in W4A16 (GPTQ format). The model has been created, tested, and evaluated by The Kaitchup. The model is NOT compatible with vLLM (as of v0.11).
Subscribe to The Kaitchup. This helps me a lot to continue quantizing and evaluating models for free. Or you prefer to give some GPU hours, "buy me a coffee"
Base model
Qwen/Qwen3-VL-8B-Instruct