Plan for much smaller model (quantized model)?

#8
by ydmhmhm - opened

Hello! I'm interested in trying this model, but it's too big to run. Do you guys have a plan for quantizing the model for consumer GPUs? Something for like 16BGB VRAM? Thanks!

Sign up or log in to comment