theprint-12B-MoE-3A

An experimental Mixture of Experts (MoE) model combining four specialized Llama 3.2 3B fine-tunes into a single 12B parameter model with ~3B active parameters per token.

For more information, please see theprint/theprint-12B-MoE-3A.

Downloads last month
291
GGUF
Model size
10B params
Architecture
llama
Hardware compatibility
Log In to view the estimation

2-bit

3-bit

4-bit

5-bit

6-bit

8-bit

16-bit

Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support

Model tree for theprint/theprint-12B-MoE-3A-GGUF

Quantized
(3)
this model