Dataset Viewer
Auto-converted to Parquet Duplicate
workload_id
string
request_count
int64
rate_per_second
float64
llm_ratio
float64
embedding_ratio
float64
arrival_pattern
string
llm_model_distribution
dict
llm_slo_deadlines
dict
embedding_model
string
embedding_batch_sizes
list
embedding_slo_deadline_ms
int64
priority_distribution
dict
hybrid_balanced
1,000
100
0.5
0.5
poisson
{ "Qwen/Qwen2.5-7B-Instruct": 0.6, "meta-llama/Llama-3.1-8B-Instruct": 0.4, "mistralai/Mistral-7B-Instruct-v0.3": null }
{ "HIGH": 500, "LOW": 2000, "NORMAL": 1000 }
BAAI/bge-m3
[ 1, 8, 16, 32 ]
200
{ "HIGH": 0.2, "LOW": 0.2, "NORMAL": 0.6 }
hybrid_burst
2,000
50
0.5
0.5
burst
{ "Qwen/Qwen2.5-7B-Instruct": 0.5, "meta-llama/Llama-3.1-8B-Instruct": 0.5, "mistralai/Mistral-7B-Instruct-v0.3": null }
{ "HIGH": 600, "LOW": 2500, "NORMAL": 1200 }
BAAI/bge-m3
[ 1, 8, 16, 32 ]
250
{ "HIGH": 0.25, "LOW": 0.25, "NORMAL": 0.5 }
hybrid_embed_heavy
1,000
150
0.2
0.8
poisson
{ "Qwen/Qwen2.5-7B-Instruct": 1, "meta-llama/Llama-3.1-8B-Instruct": null, "mistralai/Mistral-7B-Instruct-v0.3": null }
{ "HIGH": 500, "LOW": 2000, "NORMAL": 1000 }
BAAI/bge-m3
[ 8, 16, 32, 64 ]
100
{ "HIGH": 0.1, "LOW": 0.2, "NORMAL": 0.7 }
hybrid_llm_heavy
1,000
100
0.8
0.2
poisson
{ "Qwen/Qwen2.5-7B-Instruct": 0.5, "meta-llama/Llama-3.1-8B-Instruct": 0.3, "mistralai/Mistral-7B-Instruct-v0.3": 0.2 }
{ "HIGH": 400, "LOW": 1500, "NORMAL": 800 }
BAAI/bge-m3
[ 1, 8, 16 ]
150
{ "HIGH": 0.3, "LOW": 0.2, "NORMAL": 0.5 }
README.md exists but content is empty.
Downloads last month
93