Text Generation
Transformers
Safetensors
qwen3_moe
programming
code generation
code
codeqwen
Mixture of Experts
coding
coder
qwen2
chat
qwen
qwen-coder
Qwen3-Coder-30B-A3B-Instruct
Qwen3-30B-A3B
mixture of experts
128 experts
8 active experts
1 million context
qwen3
finetune
brainstorm 20x
brainstorm
optional thinking
conversational
are u wizard
#2
by
ianncity
- opened
how does this work does it merge a 14b model with the 30b model? and does the 14b model have a 1m context?
I hope you didint explain it in the description because I tried to read it all π
Hey;
See the section on "Brainstorm" ; it is a specialized adapter by yours truly.
David
See the section on "Brainstorm" ; it is a specialized adapter by yours truly.
Can you elaborate on what Brainstorm does or how it works? I've seen several models (Brainstorm x20) and i hadn't seen anything particularly different about them.
how does this work does it merge a 14b model with the 30b model? and does the 14b model have a 1m context?
It is possible because it's a MoE