KYO30 commited on
Commit
5886019
ยท
verified ยท
1 Parent(s): e8e282a

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +8 -11
app.py CHANGED
@@ -2,32 +2,29 @@ import gradio as gr
2
  from transformers import AutoModelForCausalLM, AutoTokenizer, TextIteratorStreamer
3
  import torch
4
  from threading import Thread
5
- import os # 1. Secret์„ ์ฝ๊ธฐ ์œ„ํ•ด os ๋ชจ๋“ˆ ์ž„ํฌํŠธ
6
 
7
  # --- 1. ๋ชจ๋ธ ๋กœ๋“œ (Space์˜ GPU ํ™œ์šฉ) ---
8
- MODEL_NAME = "kakaocorp/kanana-1.5-2.1b-instruct-2405"
9
-
10
- # 2. Space Setting์— ๋“ฑ๋กํ•œ Secret(HF_TOKEN)์„ ํ™˜๊ฒฝ ๋ณ€์ˆ˜์—์„œ ๊ฐ€์ ธ์˜ต๋‹ˆ๋‹ค.
11
- HF_AUTH_TOKEN = os.environ.get("HF_TOKEN")
12
 
13
  print(f"๋ชจ๋ธ์„ ๋กœ๋”ฉ ์ค‘์ž…๋‹ˆ๋‹ค: {MODEL_NAME} (Space GPU ์‚ฌ์šฉ)")
14
 
15
- # 3. ํ† ํฐ์„ ์‚ฌ์šฉํ•˜์—ฌ ์ธ์ฆ๋œ ์ƒํƒœ๋กœ ๋ชจ๋ธ์„ ๋‹ค์šด๋กœ๋“œํ•ฉ๋‹ˆ๋‹ค.
16
  tokenizer = AutoTokenizer.from_pretrained(
17
- MODEL_NAME,
18
- token=HF_AUTH_TOKEN # ์ธ์ฆ ํ† ํฐ ์ „๋‹ฌ
19
  )
20
  model = AutoModelForCausalLM.from_pretrained(
21
  MODEL_NAME,
22
  torch_dtype=torch.float16,
23
- device_map="auto",
24
- token=HF_AUTH_TOKEN # ์ธ์ฆ ํ† ํฐ ์ „๋‹ฌ
25
  )
26
  print("๋ชจ๋ธ ๋กœ๋”ฉ ์™„๋ฃŒ!")
27
 
28
  # --- 2. ์ฑ—๋ด‡ ์‘๋‹ต ํ•จ์ˆ˜ (Gradio๊ฐ€ ์ด ํ•จ์ˆ˜๋ฅผ ํ˜ธ์ถœ) ---
29
  def predict(message, history):
30
 
 
31
  history_prompt = ""
32
  for user_msg, assistant_msg in history:
33
  history_prompt += f"<bos>user\n{user_msg}\n<eos>assistant\n{assistant_msg}\n"
@@ -59,7 +56,7 @@ def predict(message, history):
59
  # --- 3. Gradio ์ฑ—๋ด‡ UI ์ƒ์„ฑ ---
60
  chatbot_ui = gr.ChatInterface(
61
  fn=predict,
62
- title="Kanana 1.5 ์ฑ—๋ด‡ ํ…Œ์ŠคํŠธ ๐Ÿค–",
63
  description=f"{MODEL_NAME} ๋ชจ๋ธ์„ ํ…Œ์ŠคํŠธํ•ฉ๋‹ˆ๋‹ค.",
64
  theme="soft",
65
  examples=[["ํ•œ๊ตญ์˜ ์ˆ˜๋„๋Š” ์–ด๋””์•ผ?"], ["AI์— ๋Œ€ํ•ด 3์ค„๋กœ ์š”์•ฝํ•ด์ค˜."]]
 
2
  from transformers import AutoModelForCausalLM, AutoTokenizer, TextIteratorStreamer
3
  import torch
4
  from threading import Thread
5
+ import os # os ๋ชจ๋“ˆ์€ ํ˜น์‹œ ๋ชจ๋ฅด๋‹ˆ ๋‚จ๊ฒจ๋‘ก๋‹ˆ๋‹ค.
6
 
7
  # --- 1. ๋ชจ๋ธ ๋กœ๋“œ (Space์˜ GPU ํ™œ์šฉ) ---
8
+ # ๋ณ„๋„ ์ธ์ฆ(ํ† ํฐ)์ด ํ•„์š” ์—†๋Š” 'kanana-nano' ๊ณต๊ฐœ ๋ชจ๋ธ๋กœ ๋ณ€๊ฒฝํ–ˆ์Šต๋‹ˆ๋‹ค.
9
+ MODEL_NAME = "kakaocorp/kanana-nano-2.1b-instruct"
 
 
10
 
11
  print(f"๋ชจ๋ธ์„ ๋กœ๋”ฉ ์ค‘์ž…๋‹ˆ๋‹ค: {MODEL_NAME} (Space GPU ์‚ฌ์šฉ)")
12
 
13
+ # ์ด ๋ชจ๋ธ์€ ๊ณต๊ฐœ ๋ชจ๋ธ์ด๋ฏ€๋กœ ํ† ํฐ(token=...) ํŒŒ๋ผ๋ฏธํ„ฐ๊ฐ€ ํ•„์š” ์—†์Šต๋‹ˆ๋‹ค.
14
  tokenizer = AutoTokenizer.from_pretrained(
15
+ MODEL_NAME
 
16
  )
17
  model = AutoModelForCausalLM.from_pretrained(
18
  MODEL_NAME,
19
  torch_dtype=torch.float16,
20
+ device_map="auto"
 
21
  )
22
  print("๋ชจ๋ธ ๋กœ๋”ฉ ์™„๋ฃŒ!")
23
 
24
  # --- 2. ์ฑ—๋ด‡ ์‘๋‹ต ํ•จ์ˆ˜ (Gradio๊ฐ€ ์ด ํ•จ์ˆ˜๋ฅผ ํ˜ธ์ถœ) ---
25
  def predict(message, history):
26
 
27
+ # Kanana์˜ ํ”„๋กฌํ”„ํŠธ ํ˜•์‹
28
  history_prompt = ""
29
  for user_msg, assistant_msg in history:
30
  history_prompt += f"<bos>user\n{user_msg}\n<eos>assistant\n{assistant_msg}\n"
 
56
  # --- 3. Gradio ์ฑ—๋ด‡ UI ์ƒ์„ฑ ---
57
  chatbot_ui = gr.ChatInterface(
58
  fn=predict,
59
+ title="Kanana-Nano ์ฑ—๋ด‡ ํ…Œ์ŠคํŠธ ๐Ÿค–",
60
  description=f"{MODEL_NAME} ๋ชจ๋ธ์„ ํ…Œ์ŠคํŠธํ•ฉ๋‹ˆ๋‹ค.",
61
  theme="soft",
62
  examples=[["ํ•œ๊ตญ์˜ ์ˆ˜๋„๋Š” ์–ด๋””์•ผ?"], ["AI์— ๋Œ€ํ•ด 3์ค„๋กœ ์š”์•ฝํ•ด์ค˜."]]