Falcon-H1: A Family of Hybrid-Head Language Models Redefining Efficiency and Performance Paper • 2507.22448 • Published Jul 30 • 68
Sparse Auto-Encoders (SAEs) for Mechanistic Interpretability Collection A compilation of sparse auto-encoders trained on large language models. • 37 items • Updated 14 days ago • 17
⚛️ Liquid Nanos Collection Library of task-specific models: https://www.liquid.ai/blog/introducing-liquid-nanos-frontier-grade-performance-on-everyday-devices • 22 items • Updated 1 day ago • 101
R-HORIZON Collection The training and evaluation datasets for Paper "How Far Can Your Large Reasoning Model Really Go in Breadth and Depth?" • 6 items • Updated Oct 22 • 7
Ministral 3 Collection A collection of edge models, with Base, Instruct and Reasoning variants, in 3 different sizes: 3B, 8B and 14B. All with vision capabilities. • 9 items • Updated 28 days ago • 133
Cerebras REAP Collection Sparse MoE models compressed using REAP (Router-weighted Expert Activation Pruning) method • 19 items • Updated 11 days ago • 72
NextCoder Collection NextCoder family of code-editing LMs developed with Selective Knowledge Transfer and its training data. • 6 items • Updated Jul 9 • 74
ACE Collection Ai2 Climate Emulator (ACE) is a family of fast ML models that simulate global atmospheric variability over time scales ranging from hours to centuries • 8 items • Updated 7 days ago • 10
SUTRA: Scalable Multilingual Language Model Architecture Paper • 2405.06694 • Published May 7, 2024 • 38