ATLAS-8B-Thinking / README.md
Jarrodbarnes's picture
Update README.md
98d57cb verified
|
raw
history blame
6.8 kB
---
license: apache-2.0
language:
- en
tags:
- reinforcement-learning
- teacher-student
- adaptive-learning
- pedagogy
- rlhf
- rlaif
base_model: Qwen/Qwen3-8B
model_type: qwen3
datasets:
- Arc-Intelligence/Arc-ATLAS-Teach-v0
model-index:
- name: ATLAS-8B-Thinking
results:
- task:
type: text-generation
name: Reinforcement Learning Teaching
dataset:
name: Arc-Intelligence/Arc-ATLAS-Teach-v0
type: Arc-Intelligence/Arc-ATLAS-Teach-v0
metrics:
- name: Non-Degradation Rate
value: 97%
type: non_degradation_rate
- name: Average Accuracy Improvement
value: +15.7%
type: average_accuracy_improvement
- name: Task Completion Rate Improvement
value: +31.2%
type: task_completion_rate_improvement
- name: Response Token Reduction
value: '-37.2%'
type: response_token_reduction
pipeline_tag: text-generation
library_name: transformers
---
# ATLAS-8B-Thinking
![ATLAS Banner](https://huggingface.co/Arc-Intelligence/ATLAS-8B-Thinking/resolve/main/ATLAS.png)
**ATLAS-8B-Thinking** is a specialized teacher model developed by Arc Intelligence, designed to solve the core reliability problem in reinforcement learning for LLMs. Standard RL fine-tuning is often brittle, leading to performance degradation where new skills are learned at the expense of old ones.
This model reframes the training process as one of **effective pedagogy**. Instead of just optimizing a student model, `ATLAS-8B-Thinking` first uses a lightweight **diagnostic probe** to assess the student's reasoning. Based on this diagnosis, it provides **adaptive guidance**—comprehensive help for struggling models and minimal intervention for capable ones. This "do no harm" approach ensures consistent capability improvement without the usual side effects of RL.
This model is a core component of the open-source [ATLAS Framework](https://github.com/Arc-Computer/ATLAS) and is designed to train and improve other language models.
## Model Performance
![Performance Chart](https://huggingface.co/Arc-Intelligence/ATLAS-8B-Thinking/resolve/main/performance-chart.png)
The ATLAS framework, using this teacher model, produces the following improvements in a student model (Qwen3-4B) compared to the student baseline. The results highlight a rare combination of increased performance, higher efficiency, and fundamental reliability.
| Metric | Improvement | Notes |
| ---------------------- | ----------- | ---------------------------------------------------------- |
| **Non-Degradation Rate** | **97%** | Core metric showing reliability and avoidance of skill loss. |
| Average Accuracy | +15.7% | Across the Arc-ATLAS-Teach-v0 evaluation set. |
| Task Completion Rate | +31.2% | Student model completes tasks it previously failed. |
| Response Tokens | -37.2% | More efficient and concise reasoning. |
## How to Use
`ATLAS-8B-Thinking` is not a standard instruction-tuned model for direct chat. It is a core component of the ATLAS training framework, designed to interact with a "student" model in a two-pass process.
### Conceptual Usage
The following is a simplified, conceptual example of the ATLAS interaction loop. The full implementation is available in the official repository.
```python
# A conceptual example of the ATLAS interaction loop
from transformers import AutoModelForCausalLM, AutoTokenizer
# Load the teacher and a student model
teacher_model = AutoModelForCausalLM.from_pretrained("Arc-Intelligence/ATLAS-8B-Thinking")
student_model = AutoModelForCausalLM.from_pretrained("Qwen/Qwen3-4B") # The model to be improved
problem = "A farmer has 52 trees planted in a row over a length of 1850 meters. What is the distance between each tree?"
# 1. Teacher creates a diagnostic probe to assess the student's initial approach
# This step is abstracted in the actual framework
diagnostic_probe = "To find the distance between the trees, what is the first critical calculation you would make?"
# 2. Student responds to the probe
# (Implementation detail: you would get the student's response here)
student_reasoning_trace = "I would divide the total length (1850m) by the number of trees (52)."
# 3. Teacher assesses the trace and provides adaptive guidance
# The teacher recognizes this common off-by-one error.
# (Implementation detail: the teacher model generates this guidance)
adaptive_guidance = "Your approach is close. Remember that 52 trees create 51 intervals between them. The distance is uniform across these intervals."
# 4. The student uses the guidance to solve the problem
final_prompt = problem + "\n" + adaptive_guidance
# (Implementation detail: the student model generates the final answer)
final_answer = "1850 meters / 51 intervals = 36.27 meters per interval."
```
### Running the Full Training Pipeline
To replicate our results or train your own models using the ATLAS framework, clone the official repository and follow the setup instructions.
```bash
# 1. Clone the repository
git clone [https://github.com/Arc-Computer/ATLAS](https://github.com/Arc-Computer/ATLAS)
cd ATLAS
# 2. Install dependencies
bash scripts/install_py312.sh
# 3. Run training
# Phase 1: Supervised Fine-Tuning (SFT)
scripts/launch.sh 4 configs/run/teacher_sft.yaml
# Phase 2: Reinforcement Learning (RL)
scripts/launch_with_server.sh 1 3 configs/run/teacher_rcl.yaml
```
## Training Details
- **Base Model:** [Qwen/Qwen3-8B](https://huggingface.co/Qwen/Qwen3-8B)
- **Training Framework:** ATLAS (SFT → RL with GRPO)
- **Key Feature:** The RL phase uses an asymmetric reward function that heavily penalizes any instance of student performance degradation, which is key to the framework's reliability.
- **Dataset:** [Arc-Intelligence/Arc-ATLAS-Teach-v0](https://huggingface.co/datasets/Arc-Intelligence/Arc-ATLAS-Teach-v0)
- **Context Length:** 8192 tokens
## Citation
If you use the ATLAS framework or our models in your research, please cite our work:
```bibtex
@misc{barnes2025atlas,
title={{ATLAS: Adaptive Teaching and Learning Alignment System for Reinforcement Learning}},
author={Jarrod Barnes and Aman Jaglan},
year={2025},
publisher={Arc Intelligence},
note={Technical Report},
url={[https://github.com/Arc-Computer/ATLAS](https://github.com/Arc-Computer/ATLAS)}
}
```
## Project Resources
- **GitHub Repository:** [https://github.com/Arc-Computer/ATLAS](https://github.com/Arc-Computer/ATLAS)
- **Companion Model:** [ATLAS-8B-Instruct](https://huggingface.co/Arc-Intelligence/ATLAS-8B-Instruct)
- **Training Dataset:** [Arc-ATLAS-Teach-v0](https://huggingface.co/datasets/Arc-Intelligence/Arc-ATLAS-Teach-v0)