Post
1524
π― Introducing Chayan: A Calibrated 4-Model LLM Router Achieving 69% Accuracy on RouterArena
We're excited to share Chayan, a cost-efficient LLM router that intelligently routes queries between 4 models to maximize accuracy while minimizing cost. Chayan just submitted to the RouterArena leaderboard and achieved 69.05% accuracy on the benchmark!
π Model: adaptive-classifier/chayan
π Dataset: RouteWorks/RouterArena
π Performance Highlights
Chayan achieves impressive results on the RouterArena benchmark:
β’ 69.05% accuracy (would rank #1 on current leaderboard)
β’ $0.333 per 1K queries
β’ +12.07pp improvement over all-mini baseline (56.98%)
β’ 99% of perfect 2-model oracle performance at 57% lower cost
Compared to our previous 2-model router (61.43% accuracy), Chayan delivers +7.62pp improvement through smarter 4-model routing.
π§ How It Works
Chayan uses an Adaptive K-NN classifier with prototype memory to route between 4 models:
β’ openai/gpt-4o-mini (fast & cheap)
β’ google/gemini-2.5-flash-lite (balanced)
β’ google/gemini-2.5-flash (capable)
β’ openai/gpt-4o (most powerful)
π Getting Started
You can use Chayan directly from HuggingFace:
from adaptive_classifier import AdaptiveClassifier
Load Chayan
router = AdaptiveClassifier.load("adaptive-classifier/chayan")
Route a query
query = "What is the capital of France?"
predictions = router.predict(query, k=4)
Get top model recommendation
best_model = predictions[0][0]
print(f"Recommended model: {best_model}")
Built with the adaptive-classifier library: https://github.com/codelion/adaptive-classifier
We're excited to share Chayan, a cost-efficient LLM router that intelligently routes queries between 4 models to maximize accuracy while minimizing cost. Chayan just submitted to the RouterArena leaderboard and achieved 69.05% accuracy on the benchmark!
π Model: adaptive-classifier/chayan
π Dataset: RouteWorks/RouterArena
π Performance Highlights
Chayan achieves impressive results on the RouterArena benchmark:
β’ 69.05% accuracy (would rank #1 on current leaderboard)
β’ $0.333 per 1K queries
β’ +12.07pp improvement over all-mini baseline (56.98%)
β’ 99% of perfect 2-model oracle performance at 57% lower cost
Compared to our previous 2-model router (61.43% accuracy), Chayan delivers +7.62pp improvement through smarter 4-model routing.
π§ How It Works
Chayan uses an Adaptive K-NN classifier with prototype memory to route between 4 models:
β’ openai/gpt-4o-mini (fast & cheap)
β’ google/gemini-2.5-flash-lite (balanced)
β’ google/gemini-2.5-flash (capable)
β’ openai/gpt-4o (most powerful)
π Getting Started
You can use Chayan directly from HuggingFace:
from adaptive_classifier import AdaptiveClassifier
Load Chayan
router = AdaptiveClassifier.load("adaptive-classifier/chayan")
Route a query
query = "What is the capital of France?"
predictions = router.predict(query, k=4)
Get top model recommendation
best_model = predictions[0][0]
print(f"Recommended model: {best_model}")
Built with the adaptive-classifier library: https://github.com/codelion/adaptive-classifier