--- license: apache-2.0 datasets: - nebius/SWE-rebench-openhands-trajectories base_model: - Qwen/Qwen3-30B-A3B-Instruct-2507 pipeline_tag: text-generation library_name: transformers tags: - code - agent --- # Model Summary **SWE-rebench-openhands-Qwen3-30B-A3B** is a 30B Rejection Sampling Fine-Tuning (RFT) checkpoint derived from [Qwen/Qwen3-30B-A3B-Instruct-2507](https://huggingface.co/Qwen/Qwen3-30B-A3B-Instruct-2507), trained on the newly released [nebius/SWE-rebench-openhands-trajectories](https://huggingface.co/datasets/nebius/SWE-rebench-openhands-trajectories) dataset. Training used a maximum sequence length of 131k tokens.
| Model | Size | Maximum Number of Turns = 100 | Maximum Number of Turns = 500 | ||||||
|---|---|---|---|---|---|---|---|---|---|
| Pass@1 | Pass@5 | Pass@1 | Pass@5 | Pass@1 | Pass@5 | Pass@1 | Pass@5 | ||
| 30B scale | |||||||||
| Qwen3-30B-A3B-Instruct-2507 | 30B | 25.2 | 44.8 | 11.8 | 24.4 | 25.7 | 44.2 | 14.2 | 26.5 |
| Qwen3-Coder-30B-A3B-Instruct | 30B | 51.9 | 67.3 | 28.7 | 42.8 | 50.0 | 63.0 | 28.1 | 38.7 |
| nebius/SWE-rebench-openhands-Qwen3-30B-A3B (Ours) | 30B | 49.7 (+24.5) |
65.4 (+20.6) |
28.1 (+16.3) |
38.7 (+14.3) |
50.3 (+24.6) |
68.3 (+24.1) |
28.1 (+13.9) |
38.7 (+12.2) |
| 100B+ scale | |||||||||
| GLM-4.5-Air | 106B | 58.2 | 73.5 | 33.8 | 42.8 | - | - | - | - |
| 200B+ scale | |||||||||
| Qwen3-235B-A22B-Instruct-2507 | 235B | 45.2 | 65.9 | 29.3 | 44.8 | 46.2 | 67.5 | 25.3 | 40.8 |
| nebius/SWE-rebench-openhands-Qwen3-235B-A22B (Ours) | 235B | 59.9 (+14.7) |
73.9 (+8.0) |
35.1 (+5.8) |
46.9 (+2.1) |
61.7 (+15.5) |
74.3 (+6.8) |
34.2 (+8.9) |
44.8 (+4.0) |
| 300B+ scale | |||||||||
| GLM-4.5 | 355B | 64.4 | 76.2 | 33.8 | 44.8 | - | - | - | - |
| Qwen3-Coder-480B-A35B-Instruct | 480B | 64.7 | 75.8 | 36.3 | 44.8 | 66.5 | 77.8 | 35.5 | 42.8 |