caizhi1 commited on
Commit
1b5976d
·
verified ·
1 Parent(s): 435a572

Update README.md

Browse files

![Linear_Attention_0926 (7)](https://cdn-uploads.huggingface.co/production/uploads/68d20104a6f8ea66da0cb447/PHRg8ipzJtr0p6sojAa5T.png)

Files changed (1) hide show
  1. README.md +2 -3
README.md CHANGED
@@ -25,7 +25,7 @@ This model continues to employ a hybrid architecture that combines linear attent
25
 
26
  <div style="display: flex; justify-content: center;">
27
  <div style="text-align: center;">
28
- <img src="https://mdn.alipayobjects.com/huamei_jcuiuk/afts/img/G_dfTZekiC4AAAAAVIAAAAgADr6CAQFr/original" width="600">
29
  <p style="margin-top: 8px; font-size: 14px;"><strong>Figure 1:</strong> Hybrid Linear Model Architecture</p>
30
  </div>
31
  </div>
@@ -97,8 +97,7 @@ for prompt in prompts:
97
  text = tokenizer.apply_chat_template(
98
  messages,
99
  tokenize=False,
100
- add_generation_prompt=True,
101
- enable_thinking=True
102
  )
103
  input_texts.append(text)
104
 
 
25
 
26
  <div style="display: flex; justify-content: center;">
27
  <div style="text-align: center;">
28
+ <img src="https://cdn-uploads.huggingface.co/production/uploads/68d20104a6f8ea66da0cb447/PHRg8ipzJtr0p6sojAa5T.png" width="1000">
29
  <p style="margin-top: 8px; font-size: 14px;"><strong>Figure 1:</strong> Hybrid Linear Model Architecture</p>
30
  </div>
31
  </div>
 
97
  text = tokenizer.apply_chat_template(
98
  messages,
99
  tokenize=False,
100
+ add_generation_prompt=True
 
101
  )
102
  input_texts.append(text)
103