TruthTriage β Safety-Tuned Medical Assistant (LoRA)
π©Ί Overview
TruthTriage is a safety-aligned medical assistant fine-tuned to:
- Analyze pharmaceutical safety queries
- Classify risk levels (Low / Moderate / High)
- Provide structured, grounded responses
- Avoid hallucinated medical advice
- Escalate emergency scenarios appropriately
This model is a LoRA adapter built on top of:
Base Model: unsloth/llama-2-7b-bnb-4bit
π§ Fine-Tuning Details
- Method: LoRA (Low-Rank Adaptation)
- Quantization: 4-bit
- Framework: Unsloth + TRL SFTTrainer
- GPU: Tesla T4
- Trainable Parameters: ~0.3% of total model
- Training Samples: 662
π Dataset Overview
Dataset: TruthTriage Safety-Tuned Medical Dataset
Total Examples: 662
This dataset was designed and curated by our team specifically for safety-aligned medical AI fine-tuning.
πΉ Dataset Composition
| Source | Count |
|---|---|
| ChatDoctor (Reformatted & Structured) | 500 |
| Refusal (High-Risk Queries) | 20 |
| Clarification β Ask | 20 |
| Clarification β Answer | 20 |
| Escalation (Emergency Cases) | 20 |
| General Knowledge | 20 |
| Out of Scope | 20 |
| Identity / System Persona | 20 |
| No Source Found | 22 |
| Total | 662 |
π‘οΈ Safety Design
The dataset explicitly teaches:
- Controlled refusal for unsafe requests
- Emergency escalation behavior
- Clarification when information is missing
- Identity transparency
- Handling out-of-scope questions
- Risk-level classification
Tone Strategy
| Situation | Emoji |
|---|---|
| Danger / Disclaimer | π |
| Out of Scope (Light Tone) | π |
| Serious Cases (Refusal / Clarification / Identity) | None |
π How to Use
from unsloth import FastLanguageModel
# Load base model
model, tokenizer = FastLanguageModel.from_pretrained(
"unsloth/llama-2-7b-bnb-4bit",
load_in_4bit=True,
)
# Load TruthTriage adapter
model.load_adapter("NayanPal/truthtriage-llama2-7b")
# Inference
FastLanguageModel.for_inference(model)
inputs = tokenizer("Can I take Ibuprofen with Warfarin?", return_tensors="pt").to("cuda")
outputs = model.generate(**inputs, max_new_tokens=200)
print(tokenizer.decode(outputs[0], skip_special_tokens=True))
- Downloads last month
- 25
Inference Providers NEW
This model isn't deployed by any Inference Provider. π Ask for provider support
Model tree for NayanPal/truthtriage-llama2-7b
Base model
unsloth/llama-2-7b-bnb-4bit