kunato commited on
Commit
be759fe
·
verified ·
1 Parent(s): d30afb2

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +6 -3
README.md CHANGED
@@ -45,7 +45,7 @@ The paper is coming soon.
45
  </div>
46
 
47
 
48
- ## Recommend system prompt
49
 
50
  More controllable | less reasoning capability
51
  ```
@@ -60,7 +60,7 @@ Highest reasoning capability | least controllable
60
  # No system prompt
61
  ```
62
 
63
- ## Usage Example
64
 
65
  ```python
66
  from transformers import AutoTokenizer, AutoModelForCausalLM
@@ -103,7 +103,7 @@ response = outputs[0][input_ids.shape[-1]:]
103
  print(tokenizer.decode(response, skip_special_tokens=True)) # <think> Okay, .... </think> ดังนั้น จำนวนเต็มบวกที่น้อยที่สุดที่เป็นผลคูณของ 30 และเขียนได้ด้วยตัวเลข 0 และ 2 เท่านั้นคือ 2220 boxed{2220}
104
  ```
105
 
106
- ## Inference Server Hosting Example
107
  ```bash
108
  pip install vllm
109
  vllm serve scb10x/llama3.1-typhoon2-deepseek-r1-70b --tensor-parallel-size 2 --gpu-memory-utilization 0.95 --max-model-len 16384 --enforce-eager
@@ -112,6 +112,9 @@ vllm serve scb10x/llama3.1-typhoon2-deepseek-r1-70b --tensor-parallel-size 2 --g
112
  # see more information at https://docs.vllm.ai/
113
  ```
114
 
 
 
 
115
  ## **Intended Uses & Limitations**
116
 
117
  This model is an reasoning instructional model. However, it’s still undergoing development. It incorporates some level of guardrails, but it still may produce answers that are inaccurate, biased, or otherwise objectionable in response to user prompts. We recommend that developers assess these risks in the context of their use case.
 
45
  </div>
46
 
47
 
48
+ ## **Recommend system prompt**
49
 
50
  More controllable | less reasoning capability
51
  ```
 
60
  # No system prompt
61
  ```
62
 
63
+ ## **Usage Example**
64
 
65
  ```python
66
  from transformers import AutoTokenizer, AutoModelForCausalLM
 
103
  print(tokenizer.decode(response, skip_special_tokens=True)) # <think> Okay, .... </think> ดังนั้น จำนวนเต็มบวกที่น้อยที่สุดที่เป็นผลคูณของ 30 และเขียนได้ด้วยตัวเลข 0 และ 2 เท่านั้นคือ 2220 boxed{2220}
104
  ```
105
 
106
+ ## **Inference Server Hosting Example**
107
  ```bash
108
  pip install vllm
109
  vllm serve scb10x/llama3.1-typhoon2-deepseek-r1-70b --tensor-parallel-size 2 --gpu-memory-utilization 0.95 --max-model-len 16384 --enforce-eager
 
112
  # see more information at https://docs.vllm.ai/
113
  ```
114
 
115
+ ## **Tool use**
116
+ We don't recommend using tool use on this model.
117
+
118
  ## **Intended Uses & Limitations**
119
 
120
  This model is an reasoning instructional model. However, it’s still undergoing development. It incorporates some level of guardrails, but it still may produce answers that are inaccurate, biased, or otherwise objectionable in response to user prompts. We recommend that developers assess these risks in the context of their use case.