DeepSupport Warm โค๏ธโ€๐Ÿฉน - LoRA adapter

This repository provides a LoRA adapter for DeepSupport Warm, an emotional-holding companion that offers gentle reflection and warm support without rushing into what to do next.

What it does โœจ

DeepSupport Warm is designed to help users feel held and less alone in the moment:

  • Validate and name feelings without judging
  • Stay with emotion first before problem-solving
  • Offer gentle grounding and a small next step only if you want

Quick start ๐Ÿš€

1) Install

pip install transformers peft accelerate torch

2) Load base model and LoRA adapter

import torch
from transformers import AutoTokenizer, AutoModelForCausalLM
from peft import PeftModel

base_id = "Qwen/Qwen2.5-32B-Instruct" # Base model we used. You may replace it with another compatible base model
lora_id = "Yukyin/deepsupport-warm-lora-oss"

tokenizer = AutoTokenizer.from_pretrained(base_id, trust_remote_code=True)
base = AutoModelForCausalLM.from_pretrained(
    base_id,
    torch_dtype=torch.float16,
    device_map="auto",
    trust_remote_code=True,
)

model = PeftModel.from_pretrained(base, lora_id)
model.eval()

messages = [
    {"role": "user", "content": "ๆˆ‘ๆœ€่ฟ‘ๅŽ‹ๅŠ›ๅพˆๅคง๏ผŒๆ„Ÿ่ง‰่‡ชๅทฑไธ€็›ดๅœจ่ขซๅฆๅฎšใ€‚"},
]

inputs = tokenizer.apply_chat_template(
    messages,
    tokenize=True,
    add_generation_prompt=True,
    return_tensors="pt",
).to(model.device)

with torch.no_grad():
    out = model.generate(
        inputs,
        max_new_tokens=256,
        do_sample=True,
        temperature=0.85,
        top_p=0.9,
        repetition_penalty=1.12,
        no_repeat_ngram_size=4,
    )

print(tokenizer.decode(out[0], skip_special_tokens=True))

Training data and release notes ๐Ÿ“Š

  • This OSS LoRA adapter is trained on de-identified versions of the original data.
  • The original internal LoRA adapter was trained on non-de-identified data and cannot be open-sourced at this time.
  • More details and examples are provided in the GitHub repo.

Safety and privacy โš ๏ธ

This project is intended for supportive conversation only.
It does not provide professional advice, diagnosis, or therapy. Please seek qualified professional help when needed.

License ๐Ÿ“œ

This adapter is released for noncommercial use. See the GitHub repo for the full license text and commercial licensing terms.

Citation ๐Ÿ“š

@software{deepsupport_warm_2026,
  author  = {Yuyan Chen},
  title   = {DeepSupport Warm: An emotional-holding companion for supportive dialogue},
  year    = {2026},
  version = {oss},
  url     = {\url{https://github.com/Yukyin/DeepSupport/DeepSupport_Warm}}
}

Links

Downloads last month
38
Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support

Model tree for Yukyin/deepsupport-warm-lora-oss

Base model

Qwen/Qwen2.5-32B
Adapter
(64)
this model