Hello! I'm interested in trying this model, but it's too big to run. Do you guys have a plan for quantizing the model for consumer GPUs? Something for like 16BGB VRAM? Thanks!
· Sign up or log in to comment