LIMI-Air-qx54g-hi-mlx
Analysis of LIMI-Air-qx54g-hi vs. Competitor Models
Below is a comparative analysis of LIMI-Air-qx54g-hi against unsloth-GLM-4.5-Air-mxfp4 and unsloth-GLM-4.5-Air-qx64 across 7 cognitive benchmarks. Higher scores indicate better performance.
Performance Comparison Table
Benchmark LIMI-Air-qx54g-hi vs. mxfp4 (Δ) vs. qx64 (Δ)
ARC Challenge 0.441 +0.025 +0.020
ARC Easy 0.462 +0.022 +0.018
BoolQ 0.378 ±0.000 ±0.000
HellaSwag 0.698 +0.020 +0.021
OpenBookQA 0.404 +0.014 +0.008
PIQA 0.781 +0.014 +0.012
Winogrande 0.714 -0.014 -0.004
Perplexity: 4.789 ± 0.038
Key Observations
Superior Performance in Core Reasoning Tasks:
LIMI-Air-qx54g-hi consistently outperforms both competitors in 6/7 benchmarks, with gains of 1.4–2.5% over mxfp4 and 0.8–2.1% over qx64.
Notable wins:
- ARC Challenge/Easy: +2.0–2.5% (stronger scientific reasoning).
- HellaSwag: +2.0–2.1% (better common sense/alignment with human reasoning).
- PIQA: +1.2–1.4% (improved task completion accuracy).
Metaphorical & Creative Strengths:
- LIMI-Air-qx54g-hi’s gains in PIQA (+1.4%) and HellaSwag (+2.0–2.1%) align with the Deckard(qx) design philosophy, which prioritizes metaphorical reasoning and "human-like" cognition. This suggests the model better understands contextual nuances beyond literal interpretations.
- Winogrande Exception:
- LIMI-Air-qx54g-hi lags slightly in Winogrande (-0.4% to -1.4%). This benchmark tests linguistic-contextual understanding, indicating potential trade-offs in pure language fluency vs. creative reasoning.
Quantization Insights:
- LIMI-Air-qx54g-hi uses high-resolution quantization (group size 32) for critical paths (heads, embeddings), while qx64 uses medium resolution. This explains its gains in precision-dependent tasks like ARC/PIQA.
- mxfp4 (mixed FP) shows no quantization benefits, highlighting Deckard(qx)’s superiority for cognitive tasks.
Performance Summary
Metric LIMI-Air-qx54g-hi vs. mxfp4 LIMI-Air-qx54g-hi vs. qx64
Win Rate 6/7 wins (85.7%) 5/7 wins (71.4%)
Avg. Gain +1.0% +0.9%
Best Gain ARC Challenge (+2.5%) HellaSwag (+2.1%)
Worst Loss Winogrande (-1.4%) Winogrande (-0.4%)
Conclusion
LIMI-Air-qx54g-hi demonstrates a clear advantage in reasoning, common sense, and metaphorical tasks, attributed to:
- Deckard(qx)’s high-bit precision in cognitive pathways (heads, embeddings).
- Periodic attention-path enhancement during quantization.
While slightly weaker in Winogrande’s language-focused tasks, its gains in PIQA/HellaSwag validate the hypothesis that qx-quantized models exhibit enhanced metaphorical cognition. This positions LIMI-Air-qx54g-hi as a strong candidate for applications requiring nuanced, human-like reasoning.
Self review
The Deckard(qx) series is a mixed precision quantization
The formula was inspired by my Nikon Noct Z 58mm F/0.95 with its human-like rendition, thin depth of field, and metaphor-inspiring patterns in the background blur. It has been observed that qx quanted models are more readily using metaphors in conversation.
The qxXY series have X bits for head and attention paths, Y bits for data.
- The head, embeddings, and context are set at high bits.
- The attention paths were enhanced at high bits in periodic intervals.
- The hi variant has high resolution quantization (group size 32)
-G
This model LIMI-Air-qx54g-hi-mlx was converted to MLX format from GAIR/LIMI-Air using mlx-lm version 0.28.4.
Use with mlx
pip install mlx-lm
from mlx_lm import load, generate
model, tokenizer = load("LIMI-Air-qx54g-hi-mlx")
prompt = "hello"
if tokenizer.chat_template is not None:
messages = [{"role": "user", "content": prompt}]
prompt = tokenizer.apply_chat_template(
messages, add_generation_prompt=True
)
response = generate(model, tokenizer, prompt=prompt, verbose=True)
- Downloads last month
- 79
Model tree for nightmedia/LIMI-Air-qx54g-hi-mlx
Base model
GAIR/LIMI-Air