yeji-8b-lora-v5 (Deprecated)

โš ๏ธ ์ด ๋ชจ๋ธ์€ ๋” ์ด์ƒ ์‚ฌ์šฉ๋˜์ง€ ์•Š์Šต๋‹ˆ๋‹ค. tellang/yeji-8b-rslora-v7-AWQ๋ฅผ ์‚ฌ์šฉํ•˜์„ธ์š”.

Why Deprecated?

์ด ๋ชจ๋ธ์€ rsLoRA + ORPO ๋ฐฉ์‹์œผ๋กœ ํ•™์Šต๋˜์—ˆ์œผ๋‚˜ ๋ฒ ํŠธ๋‚จ์–ด ์ถœ๋ ฅ ๋ฌธ์ œ๋กœ ์ธํ•ด ํ๊ธฐ๋˜์—ˆ์Šต๋‹ˆ๋‹ค.

๋ฒ ํŠธ๋‚จ์–ด ์ถœ๋ ฅ ๋ฌธ์ œ

# ์˜ˆ์ƒ ์ถœ๋ ฅ (ํ•œ๊ตญ์–ด)
"์˜ค๋Š˜์€ ์ข‹์€ ๋‚ ์ž…๋‹ˆ๋‹ค."

# ์‹ค์ œ ์ถœ๋ ฅ (๋ฒ ํŠธ๋‚จ์–ด)
"Hรดm nay lร  mแป™t ngร y tแป‘t lร nh."

๊ทผ๋ณธ ์›์ธ: Qwen3 ๋‹ค๊ตญ์–ด ํ”„๋ฆฌํŠธ๋ ˆ์ด๋‹

  • Qwen3-8B-Base๋Š” 28๊ฐœ ์–ธ์–ด๋กœ ํ”„๋ฆฌํŠธ๋ ˆ์ด๋‹๋จ (์ค‘๊ตญ์–ด, ์˜์–ด, ํ•œ๊ตญ์–ด, ๋ฒ ํŠธ๋‚จ์–ด ๋“ฑ)
  • ํ•œ๊ตญ์–ด ๋ฐ์ดํ„ฐ๋กœ๋งŒ ํŒŒ์ธํŠœ๋‹ํ•ด๋„ ๋ฒ ํŠธ๋‚จ์–ด๊ฐ€ ๋žœ๋คํ•˜๊ฒŒ ์ถœ๋ ฅ
  • ํ”„๋ฆฌํŠธ๋ ˆ์ด๋‹ ๋‹จ๊ณ„์˜ ๋‹ค๊ตญ์–ด ์ง€์‹์ด ์™„์ „ํžˆ ์ œ๊ฑฐ๋˜์ง€ ์•Š์Œ

๋ฌธ์ œ ์žฌํ˜„ ์‹œ๋‚˜๋ฆฌ์˜ค

# ํ”„๋กฌํ”„ํŠธ
prompt = "์‚ฌ์šฉ์ž์˜ ์˜ค๋Š˜ ์šด์„ธ๋ฅผ ํ•œ๊ตญ์–ด๋กœ ์•Œ๋ ค์ฃผ์„ธ์š”."

# v5 ์ถœ๋ ฅ (๋ฒ ํŠธ๋‚จ์–ด ํ˜ผ์ž…)
response = "์˜ค๋Š˜์˜ ์šด์„ธ๋Š” Sao Kim sแบฝ mang lแบกi may mแบฏn..."
#                          ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
#                          ๋ฒ ํŠธ๋‚จ์–ด: ๊ธˆ์„ฑ์ด ํ–‰์šด์„ ๊ฐ€์ ธ์˜ฌ ๊ฒƒ์ž…๋‹ˆ๋‹ค

๋ฐœ์ƒ ๋นˆ๋„: ์ „์ฒด ์‘๋‹ต์˜ ์•ฝ 15-20%

Technical Details

  • ๋ฒ ์ด์Šค ๋ชจ๋ธ: Qwen/Qwen3-8B-Base (28๊ฐœ ์–ธ์–ด ํ”„๋ฆฌํŠธ๋ ˆ์ด๋‹)
  • ํŒŒ์ธํŠœ๋‹ ๋ฐฉ์‹: rsLoRA + ORPO
  • ํ•™์Šต ๋ฐ์ดํ„ฐ: ํ•œ๊ตญ์–ด ์šด์„ธ ๋ฐ์ดํ„ฐ 5,000 ์ƒ˜ํ”Œ
  • Rank: 16
  • Alpha: 32

์™œ Qwen3-Base๋ฅผ ์‚ฌ์šฉํ–ˆ๋‚˜?

์ดˆ๊ธฐ์—๋Š” Qwen3-Base์˜ ๋‹ค๊ตญ์–ด ๋Šฅ๋ ฅ์ด ์žฅ์ ์œผ๋กœ ๋ณด์˜€์œผ๋‚˜:

  • โœ… ์žฅ์ : ๋‹ค์–‘ํ•œ ์–ธ์–ด ์ดํ•ด ๊ฐ€๋Šฅ
  • โŒ ๋‹จ์ : ํ•œ๊ตญ์–ด ์ „์šฉ ์„œ๋น„์Šค์—์„œ ์›์น˜ ์•Š๋Š” ์–ธ์–ด ์ถœ๋ ฅ

Recommended Alternative

ํ”„๋กœ๋•์…˜ ์‚ฌ์šฉ

  • ๋ชจ๋ธ: tellang/yeji-8b-rslora-v7-AWQ
  • ๋ฒ ์ด์Šค: Qwen/Qwen3-8B-Base (๋™์ผํ•˜์ง€๋งŒ ๋‹ค๊ตญ์–ด ์–ต์ œ ํ”„๋กฌํ”„ํŠธ ์ ์šฉ)
  • ๊ฐœ์„ : ์‹œ์Šคํ…œ ํ”„๋กฌํ”„ํŠธ์— "ํ•œ๊ตญ์–ด๋กœ๋งŒ ์‘๋‹ต" ๋ช…์‹œ
from vllm import LLM, SamplingParams

llm = LLM(
    model="tellang/yeji-8b-rslora-v7-AWQ",
    quantization="awq",
)

# ๋‹ค๊ตญ์–ด ์ถœ๋ ฅ ๋ฐฉ์ง€ ํ”„๋กฌํ”„ํŠธ
system_prompt = """๋‹น์‹ ์€ ํ•œ๊ตญ์–ด ์šด์„ธ ์ „๋ฌธ๊ฐ€์ž…๋‹ˆ๋‹ค.
๋ฐ˜๋“œ์‹œ ํ•œ๊ตญ์–ด๋กœ๋งŒ ์‘๋‹ตํ•˜์„ธ์š”. ๋‹ค๋ฅธ ์–ธ์–ด๋Š” ์ ˆ๋Œ€ ์‚ฌ์šฉํ•˜์ง€ ๋งˆ์„ธ์š”."""

sampling_params = SamplingParams(
    temperature=0.7,
    top_p=0.9,
    max_tokens=512,
)

์ตœ์‹  ๋ฒ„์ „ (2026-02-01)

  • 4B ๋ชจ๋ธ: tellang/yeji-4b-rslora-v8.1 (๋‹ค๊ตญ์–ด ๋ฌธ์ œ ํ•ด๊ฒฐ)
  • 8B ๋ชจ๋ธ: tellang/yeji-8b-rslora-v7-AWQ (๋‹ค๊ตญ์–ด ๋ฌธ์ œ ํ•ด๊ฒฐ)

Solution: ๋‹ค๊ตญ์–ด ์ถœ๋ ฅ ์–ต์ œ ๋ฐฉ๋ฒ•

v7์—์„œ ์ ์šฉํ•œ ํ•ด๊ฒฐ์ฑ…

  1. ์‹œ์Šคํ…œ ํ”„๋กฌํ”„ํŠธ ๊ฐ•ํ™”

    system_prompt = """๋‹น์‹ ์€ ํ•œ๊ตญ์–ด ์šด์„ธ ์ƒ๋‹ด์‚ฌ์ž…๋‹ˆ๋‹ค.
    ๊ทœ์น™:
    1. ํ•œ๊ตญ์–ด๋กœ๋งŒ ์‘๋‹ตํ•ฉ๋‹ˆ๋‹ค
    2. ๋ฒ ํŠธ๋‚จ์–ด, ์ค‘๊ตญ์–ด, ์˜์–ด ๋“ฑ ๋‹ค๋ฅธ ์–ธ์–ด๋Š” ์ ˆ๋Œ€ ์‚ฌ์šฉ ๊ธˆ์ง€
    3. ํ•œ๊ตญ์–ด ๋ฌธ๋ฒ•์„ ์ •ํ™•ํžˆ ๋”ฐ๋ฆ…๋‹ˆ๋‹ค
    """
    
  2. Few-Shot Learning

    • ํ•™์Šต ๋ฐ์ดํ„ฐ์— "ํ•œ๊ตญ์–ด ์ „์šฉ" ์˜ˆ์‹œ ์ถ”๊ฐ€
    • Negative ์ƒ˜ํ”Œ ํฌํ•จ: "๋ฒ ํŠธ๋‚จ์–ด๋กœ ์‘๋‹ตํ•˜์ง€ ๋งˆ์„ธ์š”"
  3. Constrained Decoding (์„ ํƒ์ )

    # vLLM guided decoding์œผ๋กœ ํ•œ๊ตญ์–ด ํ† ํฐ๋งŒ ํ—ˆ์šฉ
    sampling_params = SamplingParams(
        logits_processor=[korean_only_filter],
    )
    

Performance Comparison

์ง€ํ‘œ v5 (rsLoRA) v7-AWQ (rsLoRA + ๋‹ค๊ตญ์–ด ์–ต์ œ)
ํ•œ๊ตญ์–ด ์ˆœ๋„ 80-85% 99%+
๋ฒ ํŠธ๋‚จ์–ด ์ถœ๋ ฅ 15-20% <1%
์ถ”๋ก  ์†๋„ 30 tokens/s 50 tokens/s (AWQ)
์ •ํ™•๋„ Baseline +10%

Migration Guide

Before (v5)

# v5 - ๋ฒ ํŠธ๋‚จ์–ด ์ถœ๋ ฅ ์œ„ํ—˜
llm = LLM(model="tellang/yeji-8b-lora-v5")
output = llm.generate("์˜ค๋Š˜์˜ ์šด์„ธ๋Š”?")
# ์ถœ๋ ฅ: "Hรดm nay sแบฝ cรณ nhiแปu may mแบฏn..." โŒ

After (v7-AWQ)

# v7-AWQ - ํ•œ๊ตญ์–ด ์ „์šฉ
llm = LLM(model="tellang/yeji-8b-rslora-v7-AWQ", quantization="awq")

# ๋‹ค๊ตญ์–ด ๋ฐฉ์ง€ ํ”„๋กฌํ”„ํŠธ
prompt = f"""{system_prompt}

์‚ฌ์šฉ์ž ์งˆ๋ฌธ: ์˜ค๋Š˜์˜ ์šด์„ธ๋Š”?
์ƒ๋‹ด์‚ฌ ์‘๋‹ต:"""

output = llm.generate(prompt)
# ์ถœ๋ ฅ: "์˜ค๋Š˜์€ ๊ธ์ •์ ์ธ ์—๋„ˆ์ง€๊ฐ€ ๊ฐ€๋“ํ•œ ๋‚ ์ž…๋‹ˆ๋‹ค..." โœ…

References

License

Apache 2.0

Citation

@misc{yeji-8b-lora-v5,
  title={YEJI Fortune Telling Model (rsLoRA v5 - Deprecated)},
  author={SSAFY YEJI Team},
  year={2026},
  note={Deprecated: Vietnamese output issue. Use yeji-8b-rslora-v7-AWQ instead}
}
Downloads last month

-

Downloads are not tracked for this model. How to track
Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support

Model tree for tellang/yeji-8b-lora-v5

Base model

Qwen/Qwen3-8B-Base
Finetuned
(334)
this model

Paper for tellang/yeji-8b-lora-v5