Hugging Face
Models
Datasets
Spaces
Buckets
new
Docs
Enterprise
Pricing
Log In
Sign Up
2
5
Bambuu
bambuuai
Follow
lambdago's profile picture
hongyu05's profile picture
2 followers
ยท
15 following
bambuuai
AI & ML interests
None yet
Recent Activity
upvoted
an
article
about 14 hours ago
Introducing Storage Buckets on the Hugging Face Hub
reacted
to
SeaWolf-AI
's
post
with ๐ฅ
about 14 hours ago
๐๏ธ Smol AI WorldCup: A 4B Model Just Beat 8B โ Here's the Data We evaluated 18 small language models from 12 makers on 125 questions across 7 languages. The results challenge the assumption that bigger is always better. Community Article: https://huggingface.co/blog/FINAL-Bench/smol-worldcup Live Leaderboard: https://huggingface.co/spaces/ginigen-ai/smol-worldcup Dataset: https://huggingface.co/datasets/ginigen-ai/smol-worldcup What we found: โ Gemma-3n-E4B (4B, 2GB RAM) outscores Qwen3-8B (8B, 5.5GB). Doubling parameters gained only 0.4 points. RAM cost: 2.75x more. โ GPT-OSS-20B fits in 1.5GB yet matches Champions-league dense models requiring 8.5GB. MoE architecture is the edge AI game-changer. โ Thinking models hurt structured output. DeepSeek-R1-7B scores 8.7 points below same-size Qwen3-8B and runs 2.7x slower. โ A 1.3B model fabricates confident fake content 80% of the time when prompted with nonexistent entities. Qwen3 family hits 100% trap detection across all sizes. โ Qwen3-1.7B (1.2GB) outscores Mistral-7B, Llama-3.1-8B, and DeepSeek-R1-14B. Latest architecture at 1.7B beats older architecture at 14B. What makes this benchmark different? Most benchmarks ask "how smart?" โ we measure five axes simultaneously: Size, Honesty, Intelligence, Fast, Thrift (SHIFT). Our ranking metric WCS = sqrt(SHIFT x PIR_norm) rewards models that are both high-quality AND efficient. Smart but massive? Low rank. Tiny but poor? Also low. Top 5 by WCS: 1. GPT-OSS-20B โ WCS 82.6 โ 1.5GB โ Raspberry Pi tier 2. Gemma-3n-E4B โ WCS 81.8 โ 2.0GB โ Smartphone tier 3. Llama-4-Scout โ WCS 79.3 โ 240 tok/s โ Fastest model 4. Qwen3-4B โ WCS 76.6 โ 2.8GB โ Smartphone tier 5. Qwen3-1.7B โ WCS 76.1 โ 1.2GB โ IoT tier Built in collaboration with the FINAL Bench research team. Interoperable with ALL Bench Leaderboard for full small-to-large model comparison. Dataset is open under Apache 2.0 (125 questions, 7 languages). We welcome new model submissions.
liked
a dataset
2 days ago
zeroentropy/polysemy
View all activity
Organizations
bambuuai
's activity
All
Models
Datasets
Spaces
Papers
Collections
Community
Posts
Upvotes
Likes
Articles
liked
a dataset
2 days ago
zeroentropy/polysemy
Viewer
โข
Updated
Nov 12, 2025
โข
650
โข
13
โข
2
liked
2 models
6 days ago
fal/Qwen-Image-Edit-2511-Multiple-Angles-LoRA
Image-to-Image
โข
Updated
Jan 7
โข
59k
โข
โข
1.11k
NoesisLab/Kai-30B-Instruct
Text Generation
โข
33B
โข
Updated
7 days ago
โข
471
โข
20
liked
a model
8 days ago
Qwen/Qwen3.5-27B-GPTQ-Int4
Image-Text-to-Text
โข
28B
โข
Updated
8 days ago
โข
85.3k
โข
23
liked
a model
9 days ago
Qwen/Qwen3.5-35B-A3B
Image-Text-to-Text
โข
36B
โข
Updated
13 days ago
โข
1.35M
โข
โข
1.08k