Update tokenizer_config.json
Browse files- tokenizer_config.json +2 -2
tokenizer_config.json
CHANGED
|
@@ -238,7 +238,7 @@
|
|
| 238 |
"content": {
|
| 239 |
"type": "string"
|
| 240 |
},
|
| 241 |
-
"
|
| 242 |
"type": "string"
|
| 243 |
},
|
| 244 |
"role": {
|
|
@@ -253,7 +253,7 @@
|
|
| 253 |
}
|
| 254 |
},
|
| 255 |
"type": "object",
|
| 256 |
-
"x-regex": "^(?:(?:<think>)?\\s*(?P<
|
| 257 |
},
|
| 258 |
"split_special_tokens": false,
|
| 259 |
"tokenizer_class": "Qwen2Tokenizer",
|
|
|
|
| 238 |
"content": {
|
| 239 |
"type": "string"
|
| 240 |
},
|
| 241 |
+
"reasoning": {
|
| 242 |
"type": "string"
|
| 243 |
},
|
| 244 |
"role": {
|
|
|
|
| 253 |
}
|
| 254 |
},
|
| 255 |
"type": "object",
|
| 256 |
+
"x-regex": "^(?:(?:<think>)?\\s*(?P<reasoning>.+?)\\s*</think>)?\\s*(?:<tool_call>(?P<tool_calls>.*?)\\s*</tool_call>)?\\s*(?P<content>.+?)?\\s*$"
|
| 257 |
},
|
| 258 |
"split_special_tokens": false,
|
| 259 |
"tokenizer_class": "Qwen2Tokenizer",
|