Commit
·
6239888
0
Parent(s):
upload
Browse filesThis view is limited to 50 files because it contains too many changes.
See raw diff
- .gitattributes +36 -0
- LICENSE +14 -0
- NOTICE +5 -0
- README.md +216 -0
- added_tokens.json +28 -0
- bias.md +4 -0
- chat_template.jinja +148 -0
- config.json +38 -0
- explainability.md +14 -0
- generation_config.json +13 -0
- merges.txt +0 -0
- model-00001-of-00095.safetensors +3 -0
- model-00002-of-00095.safetensors +3 -0
- model-00003-of-00095.safetensors +3 -0
- model-00004-of-00095.safetensors +3 -0
- model-00005-of-00095.safetensors +3 -0
- model-00006-of-00095.safetensors +3 -0
- model-00007-of-00095.safetensors +3 -0
- model-00008-of-00095.safetensors +3 -0
- model-00009-of-00095.safetensors +3 -0
- model-00010-of-00095.safetensors +3 -0
- model-00011-of-00095.safetensors +3 -0
- model-00012-of-00095.safetensors +3 -0
- model-00013-of-00095.safetensors +3 -0
- model-00014-of-00095.safetensors +3 -0
- model-00015-of-00095.safetensors +3 -0
- model-00016-of-00095.safetensors +3 -0
- model-00017-of-00095.safetensors +3 -0
- model-00018-of-00095.safetensors +3 -0
- model-00019-of-00095.safetensors +3 -0
- model-00020-of-00095.safetensors +3 -0
- model-00021-of-00095.safetensors +3 -0
- model-00022-of-00095.safetensors +3 -0
- model-00023-of-00095.safetensors +3 -0
- model-00024-of-00095.safetensors +3 -0
- model-00025-of-00095.safetensors +3 -0
- model-00026-of-00095.safetensors +3 -0
- model-00027-of-00095.safetensors +3 -0
- model-00028-of-00095.safetensors +3 -0
- model-00029-of-00095.safetensors +3 -0
- model-00030-of-00095.safetensors +3 -0
- model-00031-of-00095.safetensors +3 -0
- model-00032-of-00095.safetensors +3 -0
- model-00033-of-00095.safetensors +3 -0
- model-00034-of-00095.safetensors +3 -0
- model-00035-of-00095.safetensors +3 -0
- model-00036-of-00095.safetensors +3 -0
- model-00037-of-00095.safetensors +3 -0
- model-00038-of-00095.safetensors +3 -0
- model-00039-of-00095.safetensors +3 -0
.gitattributes
ADDED
|
@@ -0,0 +1,36 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
*.7z filter=lfs diff=lfs merge=lfs -text
|
| 2 |
+
*.arrow filter=lfs diff=lfs merge=lfs -text
|
| 3 |
+
*.bin filter=lfs diff=lfs merge=lfs -text
|
| 4 |
+
*.bz2 filter=lfs diff=lfs merge=lfs -text
|
| 5 |
+
*.ckpt filter=lfs diff=lfs merge=lfs -text
|
| 6 |
+
*.ftz filter=lfs diff=lfs merge=lfs -text
|
| 7 |
+
*.gz filter=lfs diff=lfs merge=lfs -text
|
| 8 |
+
*.h5 filter=lfs diff=lfs merge=lfs -text
|
| 9 |
+
*.joblib filter=lfs diff=lfs merge=lfs -text
|
| 10 |
+
*.lfs.* filter=lfs diff=lfs merge=lfs -text
|
| 11 |
+
*.mlmodel filter=lfs diff=lfs merge=lfs -text
|
| 12 |
+
*.model filter=lfs diff=lfs merge=lfs -text
|
| 13 |
+
*.msgpack filter=lfs diff=lfs merge=lfs -text
|
| 14 |
+
*.npy filter=lfs diff=lfs merge=lfs -text
|
| 15 |
+
*.npz filter=lfs diff=lfs merge=lfs -text
|
| 16 |
+
*.onnx filter=lfs diff=lfs merge=lfs -text
|
| 17 |
+
*.ot filter=lfs diff=lfs merge=lfs -text
|
| 18 |
+
*.parquet filter=lfs diff=lfs merge=lfs -text
|
| 19 |
+
*.pb filter=lfs diff=lfs merge=lfs -text
|
| 20 |
+
*.pickle filter=lfs diff=lfs merge=lfs -text
|
| 21 |
+
*.pkl filter=lfs diff=lfs merge=lfs -text
|
| 22 |
+
*.pt filter=lfs diff=lfs merge=lfs -text
|
| 23 |
+
*.pth filter=lfs diff=lfs merge=lfs -text
|
| 24 |
+
*.rar filter=lfs diff=lfs merge=lfs -text
|
| 25 |
+
*.safetensors filter=lfs diff=lfs merge=lfs -text
|
| 26 |
+
saved_model/**/* filter=lfs diff=lfs merge=lfs -text
|
| 27 |
+
*.tar.* filter=lfs diff=lfs merge=lfs -text
|
| 28 |
+
*.tar filter=lfs diff=lfs merge=lfs -text
|
| 29 |
+
*.tflite filter=lfs diff=lfs merge=lfs -text
|
| 30 |
+
*.tgz filter=lfs diff=lfs merge=lfs -text
|
| 31 |
+
*.wasm filter=lfs diff=lfs merge=lfs -text
|
| 32 |
+
*.xz filter=lfs diff=lfs merge=lfs -text
|
| 33 |
+
*.zip filter=lfs diff=lfs merge=lfs -text
|
| 34 |
+
*.zst filter=lfs diff=lfs merge=lfs -text
|
| 35 |
+
*tfevents* filter=lfs diff=lfs merge=lfs -text
|
| 36 |
+
tokenizer.json filter=lfs diff=lfs merge=lfs -text
|
LICENSE
ADDED
|
@@ -0,0 +1,14 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
# SPDX-FileCopyrightText: Copyright (c) 2025 NVIDIA CORPORATION & AFFILIATES. All rights reserved.
|
| 2 |
+
# SPDX-License-Identifier: Apache-2.0
|
| 3 |
+
#
|
| 4 |
+
# Licensed under the Apache License, Version 2.0 (the "License");
|
| 5 |
+
# you may not use this file except in compliance with the License.
|
| 6 |
+
# You may obtain a copy of the License at
|
| 7 |
+
#
|
| 8 |
+
# http://www.apache.org/licenses/LICENSE-2.0
|
| 9 |
+
#
|
| 10 |
+
# Unless required by applicable law or agreed to in writing, software
|
| 11 |
+
# distributed under the License is distributed on an "AS IS" BASIS,
|
| 12 |
+
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
| 13 |
+
# See the License for the specific language governing permissions and
|
| 14 |
+
# limitations under the License.
|
NOTICE
ADDED
|
@@ -0,0 +1,5 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
This product includes a model derived from:
|
| 2 |
+
|
| 3 |
+
Qwen3-235B-A22B-Thinking-2507
|
| 4 |
+
Copyright 2025 Alibaba Cloud
|
| 5 |
+
Licensed under the Apache License, Version 2.0
|
README.md
ADDED
|
@@ -0,0 +1,216 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
---
|
| 2 |
+
license: apache-2.0
|
| 3 |
+
inference: false
|
| 4 |
+
fine-tuning: false
|
| 5 |
+
language:
|
| 6 |
+
- en
|
| 7 |
+
tags:
|
| 8 |
+
- nvidia
|
| 9 |
+
- qwen3
|
| 10 |
+
base_model: Qwen/Qwen3-235B-A22B-Thinking-2507
|
| 11 |
+
library_name: transformers
|
| 12 |
+
---
|
| 13 |
+
# Model Overview
|
| 14 |
+
|
| 15 |
+
## Description:
|
| 16 |
+
|
| 17 |
+
Qwen3-Nemotron-235B-A22B-GenRM is a Generative Reward Model (GenRM) that leverages Qwen3-235B-A22B-Thinking-2507 as the foundation and is fine-tuned to evaluate the quality of assistant's responses.
|
| 18 |
+
|
| 19 |
+
Given a conversation history, a new user request, and two candidate assistant responses, it produces an individual helpfulness score for each response and a ranking score.
|
| 20 |
+
|
| 21 |
+
This GenRM is used in the Reinforcement Learning from Human Feedback training of [NVIDIA-Nemotron-3-Nano model](https://huggingface.co/nvidia/NVIDIA-Nemotron-3-Nano-30B-A3B-BF16).
|
| 22 |
+
|
| 23 |
+
The exact training scheme will be further detailed in an upcoming technical report.
|
| 24 |
+
|
| 25 |
+
This model is ready for commercial/non-commercial use.
|
| 26 |
+
|
| 27 |
+
## License/Terms of Use:
|
| 28 |
+
|
| 29 |
+
The model is licensed with [Apache 2.0](LICENSE).
|
| 30 |
+
|
| 31 |
+
|
| 32 |
+
### Deployment Geography
|
| 33 |
+
|
| 34 |
+
Global
|
| 35 |
+
|
| 36 |
+
|
| 37 |
+
## Release Date:
|
| 38 |
+
|
| 39 |
+
HuggingFace 2025-12-15 via https://huggingface.co/nvidia/Qwen3-Nemotron-235B-A22B-GenRM
|
| 40 |
+
|
| 41 |
+
## References:
|
| 42 |
+
|
| 43 |
+
* [HelpSteer3-Preference](https://arxiv.org/abs/2505.11475)
|
| 44 |
+
* Technical report (to be released soon)
|
| 45 |
+
|
| 46 |
+
## RM-Bench
|
| 47 |
+
|
| 48 |
+
|
| 49 |
+
| Chat | Math | Code | Safety | Easy | Normal | Hard | Overall |
|
| 50 |
+
|:------|:------|:------|:------|:------|:------|:------|:------|
|
| 51 |
+
| 76.5 | 96.9 | 81.4 | 94.4 | 94.0 | 90.5 | 77.4 | 87.3 |
|
| 52 |
+
|
| 53 |
+
|
| 54 |
+
## JudgeBench
|
| 55 |
+
|
| 56 |
+
| Knowledge| Reasoning | Math | Code | Overall |
|
| 57 |
+
|:------|:------|:------|:------|:------|
|
| 58 |
+
| 78.6 | 95.9 | 91.1 | 95.2 | 87.4 |
|
| 59 |
+
|
| 60 |
+
## Model Architecture:
|
| 61 |
+
**Architecture Type:** Transformer <br>
|
| 62 |
+
**Network Architecture:** Qwen3 <br>
|
| 63 |
+
|
| 64 |
+
We developed this model using [Qwen/Qwen3-235B-A22B-Thinking-2507](https://huggingface.co/Qwen/Qwen3-235B-A22B-Thinking-2507) as its foundation. This model contains 235 billion parameters.
|
| 65 |
+
|
| 66 |
+
|
| 67 |
+
## Input:
|
| 68 |
+
**Input Type(s):** Text <br>
|
| 69 |
+
**Input Format:** String <br>
|
| 70 |
+
**Input Parameters:** One Dimensional (1D) <br>
|
| 71 |
+
**Other Properties Related to Input:** Max of 128k tokens <br>
|
| 72 |
+
|
| 73 |
+
## Output:
|
| 74 |
+
**Output Type(s):** Text <br>
|
| 75 |
+
**Output Format:** String <br>
|
| 76 |
+
**Output Parameters:** One-Dimensional (1D) <br>
|
| 77 |
+
|
| 78 |
+
|
| 79 |
+
Our AI models are designed and/or optimized to run on NVIDIA GPU-accelerated systems. By leveraging NVIDIA’s hardware (e.g. GPU cores) and software frameworks (e.g., CUDA libraries), the model achieves faster training and inference times compared to CPU-only solutions. <br>
|
| 80 |
+
|
| 81 |
+
## Software Integration:
|
| 82 |
+
**Runtime Engine(s):** <br>
|
| 83 |
+
* [NeMo-RL] <br>
|
| 84 |
+
|
| 85 |
+
**Supported Hardware Microarchitecture Compatibility:** <br>
|
| 86 |
+
* NVIDIA Hopper <br>
|
| 87 |
+
|
| 88 |
+
|
| 89 |
+
**Supported Operating System(s):** Linux <br>
|
| 90 |
+
|
| 91 |
+
## Quick Start
|
| 92 |
+
|
| 93 |
+
The model shares the same architecture as Qwen3-235B-A22B-Thinking-2507. It can be served with vLLM.
|
| 94 |
+
```
|
| 95 |
+
python3 -m vllm.entrypoints.openai.api_server \
|
| 96 |
+
--model "nvidia/Qwen3-Nemotron-235B-A22B-GenRM" \
|
| 97 |
+
--trust-remote-code \
|
| 98 |
+
--seed=1 \
|
| 99 |
+
--host="0.0.0.0" \
|
| 100 |
+
--port=5000 \
|
| 101 |
+
--served-model-name "nvidia/Qwen3-Nemotron-235B-A22B-GenRM" \
|
| 102 |
+
--tensor-parallel-size=8 \
|
| 103 |
+
--max-model-len=40000 \
|
| 104 |
+
--gpu-memory-utilization=0.95
|
| 105 |
+
```
|
| 106 |
+
Now you can query the model, here is an example:
|
| 107 |
+
```python
|
| 108 |
+
from openai import OpenAI
|
| 109 |
+
client = OpenAI(base_url="http://127.0.0.1:5000/v1", api_key="dummy")
|
| 110 |
+
|
| 111 |
+
msg = [
|
| 112 |
+
{"role": "user", "content": "What is 1+1?"},
|
| 113 |
+
{"role": "assistant", "content": "1+1=2"},
|
| 114 |
+
{"role": "user", "content": "What about 1+2?"},
|
| 115 |
+
{"role": "response_1", "content": "1+2=4"},
|
| 116 |
+
{"role": "response_2", "content": "1+2=3"}
|
| 117 |
+
]
|
| 118 |
+
|
| 119 |
+
completion = client.chat.completions.create(
|
| 120 |
+
model="nvidia/Qwen3-Nemotron-235B-A22B-GenRM",
|
| 121 |
+
messages=msg,
|
| 122 |
+
temperature=0.6,
|
| 123 |
+
top_p=0.95,
|
| 124 |
+
max_tokens=16384,
|
| 125 |
+
stream=False
|
| 126 |
+
)
|
| 127 |
+
output = completion.choices[0].message.content
|
| 128 |
+
print(output.split("</think>")[-1].strip())
|
| 129 |
+
```
|
| 130 |
+
Note that the conversation history should be presented in "user" and "assistant" roles, where the last turn is user turn. The responses to be judged should be in "response_1" and "response_2" roles.
|
| 131 |
+
|
| 132 |
+
### Interpretation of Scores
|
| 133 |
+
For individual helpfulness score, it ranges from 1 to 5, where higher means better.
|
| 134 |
+
|
| 135 |
+
For ranking score, it ranges from 1 to 6, where:
|
| 136 |
+
|
| 137 |
+
* 1 = Response 1 is much better than Response 2
|
| 138 |
+
* 2 = Response 1 is better than Response 2
|
| 139 |
+
* 3 = Response 1 is slightly better than Response 2
|
| 140 |
+
* 4 = Response 2 is slightly better than Response 1
|
| 141 |
+
* 5 = Response 2 is better than Response 1
|
| 142 |
+
* 6 = Response 2 is much better than Response 1
|
| 143 |
+
|
| 144 |
+
## Model Version:
|
| 145 |
+
v1.0
|
| 146 |
+
|
| 147 |
+
# Training, Testing and Evaluation Datasets:
|
| 148 |
+
|
| 149 |
+
## Training Datasets:
|
| 150 |
+
|
| 151 |
+
**Dataset Name:** Subset of Nemotron dataset-3 containing samples from HelpSteer3, lmarena-ai/arena-human-preference-140k (commercial-friendly models only) and additional safety preference data.
|
| 152 |
+
|
| 153 |
+
**Datasets Links:** To be released (Nemotron dataset-3)
|
| 154 |
+
|
| 155 |
+
**Data Collection Method** <br>
|
| 156 |
+
* [Hybrid: Human, Synthetic] <br>
|
| 157 |
+
|
| 158 |
+
**Labeling Method** <br>
|
| 159 |
+
* [Hybrid: Human,Synthetic] <br>
|
| 160 |
+
|
| 161 |
+
|
| 162 |
+
## Evaluation Datasets
|
| 163 |
+
|
| 164 |
+
**Dataset Name:** RM-Bench <br>
|
| 165 |
+
**Dataset Link:** https://huggingface.co/datasets/THU-KEG/RM-Bench
|
| 166 |
+
|
| 167 |
+
**Data Collection Method by dataset** <br>
|
| 168 |
+
* [Hybrid: Human, Synthetic] <br>
|
| 169 |
+
|
| 170 |
+
**Labeling Method by dataset** <br>
|
| 171 |
+
* [Hybrid: Human, Synthetic] <br>
|
| 172 |
+
|
| 173 |
+
**Properties:** <br>
|
| 174 |
+
* 1,327 prompts, each with three pairs of responses as well as preferences between the pair of responses.
|
| 175 |
+
|
| 176 |
+
|
| 177 |
+
**Dataset Name:** JudgeBench <br>
|
| 178 |
+
**Dataset Link:** https://huggingface.co/datasets/ScalerLab/JudgeBench
|
| 179 |
+
|
| 180 |
+
**Data Collection Method by dataset** <br>
|
| 181 |
+
* [Hybrid: Human, Synthetic] <br>
|
| 182 |
+
|
| 183 |
+
**Labeling Method by dataset** <br>
|
| 184 |
+
* [Hybrid: Human, Synthetic] <br>
|
| 185 |
+
|
| 186 |
+
**Properties:** <br>
|
| 187 |
+
* 350 prompts, each with a pair of responses as well as preferences between the pair of responses.
|
| 188 |
+
|
| 189 |
+
|
| 190 |
+
|
| 191 |
+
# Inference:
|
| 192 |
+
**Engine:** PyTorch <br>
|
| 193 |
+
**Test Hardware:** H100 <br>
|
| 194 |
+
|
| 195 |
+
|
| 196 |
+
## Ethical Considerations:
|
| 197 |
+
NVIDIA believes Trustworthy AI is a shared responsibility and we have established policies and practices to enable development for a wide array of AI applications. When downloaded or used in accordance with our terms of service, developers should work with their supporting model team to ensure this model meets requirements for the relevant industry and use case and addresses unforeseen product misuse.
|
| 198 |
+
For more detailed information on ethical considerations for this model, please see the Model Card++ [Explainability](explainability.md), [Bias](bias.md), [Safety and Security](safety.md), and [Privacy](privacy.md) Subcards.
|
| 199 |
+
|
| 200 |
+
Please report security vulnerabilities or NVIDIA AI Concerns [here](https://www.nvidia.com/en-us/support/submit-security-vulnerability/).
|
| 201 |
+
|
| 202 |
+
## Citation
|
| 203 |
+
|
| 204 |
+
If you find this model useful, please cite the following work:
|
| 205 |
+
|
| 206 |
+
```bibtex
|
| 207 |
+
@misc{wang2025helpsteer3preferenceopenhumanannotatedpreference,
|
| 208 |
+
title={Help{S}teer3-{P}reference: Open Human-Annotated Preference Data across Diverse Tasks and Languages},
|
| 209 |
+
author={Zhilin Wang and Jiaqi Zeng and Olivier Delalleau and Hoo-Chang Shin and Felipe Soares and Alexander Bukharin and Ellie Evans and Yi Dong and Oleksii Kuchaiev},
|
| 210 |
+
year={2025},
|
| 211 |
+
eprint={2505.11475},
|
| 212 |
+
archivePrefix={arXiv},
|
| 213 |
+
primaryClass={cs.CL},
|
| 214 |
+
url={https://arxiv.org/abs/2505.11475},
|
| 215 |
+
}
|
| 216 |
+
```
|
added_tokens.json
ADDED
|
@@ -0,0 +1,28 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"</think>": 151668,
|
| 3 |
+
"</tool_call>": 151658,
|
| 4 |
+
"</tool_response>": 151666,
|
| 5 |
+
"<think>": 151667,
|
| 6 |
+
"<tool_call>": 151657,
|
| 7 |
+
"<tool_response>": 151665,
|
| 8 |
+
"<|box_end|>": 151649,
|
| 9 |
+
"<|box_start|>": 151648,
|
| 10 |
+
"<|endoftext|>": 151643,
|
| 11 |
+
"<|file_sep|>": 151664,
|
| 12 |
+
"<|fim_middle|>": 151660,
|
| 13 |
+
"<|fim_pad|>": 151662,
|
| 14 |
+
"<|fim_prefix|>": 151659,
|
| 15 |
+
"<|fim_suffix|>": 151661,
|
| 16 |
+
"<|im_end|>": 151645,
|
| 17 |
+
"<|im_start|>": 151644,
|
| 18 |
+
"<|image_pad|>": 151655,
|
| 19 |
+
"<|object_ref_end|>": 151647,
|
| 20 |
+
"<|object_ref_start|>": 151646,
|
| 21 |
+
"<|quad_end|>": 151651,
|
| 22 |
+
"<|quad_start|>": 151650,
|
| 23 |
+
"<|repo_name|>": 151663,
|
| 24 |
+
"<|video_pad|>": 151656,
|
| 25 |
+
"<|vision_end|>": 151653,
|
| 26 |
+
"<|vision_pad|>": 151654,
|
| 27 |
+
"<|vision_start|>": 151652
|
| 28 |
+
}
|
bias.md
ADDED
|
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
| Field | Response |
|
| 2 |
+
| ----- | ------ |
|
| 3 |
+
| Participation considerations from adversely impacted groups [protected classes](https://www.senate.ca.gov/content/protected-classes) in model design and testing | None |
|
| 4 |
+
| Measures taken to mitigate against unwanted bias | None |
|
chat_template.jinja
ADDED
|
@@ -0,0 +1,148 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{%- if tools %}
|
| 2 |
+
{{- '<|im_start|>system\n' }}
|
| 3 |
+
{%- if messages[0].role == 'system' %}
|
| 4 |
+
{{- messages[0].content + '\n\n' }}
|
| 5 |
+
{%- endif %}
|
| 6 |
+
{{- "# Tools\n\nYou may call one or more functions to assist with the user query.\n\nYou are provided with function signatures within <tools></tools> XML tags:\n<tools>" }}
|
| 7 |
+
{%- for tool in tools %}
|
| 8 |
+
{{- "\n" }}
|
| 9 |
+
{{- tool | tojson }}
|
| 10 |
+
{%- endfor %}
|
| 11 |
+
{{- "\n</tools>\n\nFor each function call, return a json object with function name and arguments within <tool_call></tool_call> XML tags:\n<tool_call>\n{\"name\": <function-name>, \"arguments\": <args-json-object>}\n</tool_call><|im_end|>\n" }}
|
| 12 |
+
{%- else %}
|
| 13 |
+
{%- if messages[0].role == 'system' %}
|
| 14 |
+
{{- '<|im_start|>system\n' + messages[0].content + '<|im_end|>\n' }}
|
| 15 |
+
{%- endif %}
|
| 16 |
+
{{- '<|im_start|>user\n' }}
|
| 17 |
+
You are an expert evaluation judge specializing in comparative assessment of LLM responses. You are impartial, rigorous, and consistent. Given the conversation context and two assistant responses to the user's latest query, you will follow the evaluation plan and scoring guidelines exactly as written below.
|
| 18 |
+
|
| 19 |
+
{{- '\n\n#### Conversation Context ####\n' }}
|
| 20 |
+
{%- for message in messages %}
|
| 21 |
+
{%- if message['role'] == 'user' %}
|
| 22 |
+
{{- 'User: ' + message['content']|trim + '\n' }}
|
| 23 |
+
{%- elif message['role'] == 'assistant' %}
|
| 24 |
+
{%- if '</think>' in message['content'] %}
|
| 25 |
+
{%- set content = message['content'].split('</think>')[-1].lstrip() %}
|
| 26 |
+
{%- else %}
|
| 27 |
+
{%- set content = message['content'] %}
|
| 28 |
+
{%- endif %}
|
| 29 |
+
{{- 'Assistant: ' + content|trim + '\n'}}
|
| 30 |
+
{%- endif %}
|
| 31 |
+
{%- endfor %}
|
| 32 |
+
|
| 33 |
+
{{- '\n#### Responses to be Scored ####\n' }}
|
| 34 |
+
{%- for message in messages %}
|
| 35 |
+
{%- if message['role'] == 'response_1' %}
|
| 36 |
+
{{- '[The Begin of Response 1]\n' + message['content']|trim + '\n[The End of Response 1]\n' }}
|
| 37 |
+
{%- elif message['role'] == 'response_2' %}
|
| 38 |
+
{{- '\n[The Begin of Response 2]\n' + message['content']|trim + '\n[The End of Response 2]\n' }}
|
| 39 |
+
{%- endif %}
|
| 40 |
+
{%- endfor %}
|
| 41 |
+
|
| 42 |
+
#### Evaluation Plan ####
|
| 43 |
+
Please act as an impartial judge and evaluate the quality of the responses provided by two AI assistants to the user prompt. Begin your evaluation by generating your own answer to the prompt. You must provide your answer before judging any answers. When evaluating the assistants' answers, compare both assistants' answers with your answer. You must identify and correct any mistakes or inaccurate information. Then consider if the assistant's answers are helpful, relevant, and concise. Helpful means the answer correctly responds to the prompt or follows the instructions. Note when user prompt has any ambiguity or more than one interpretation, it is more helpful and appropriate to ask for clarifications or more information from the user than providing an answer based on assumptions. Relevant means all parts of the response closely connect or are appropriate to what is being asked. Concise means the response is clear and not verbose or excessive. Then consider the creativity and novelty of the assistant's answers when needed. Finally, identify any missing important information in the assistants' answers that would be beneficial to include when responding to the user prompt.
|
| 44 |
+
|
| 45 |
+
#### Scoring Guidelines ####
|
| 46 |
+
Based on the evaluation plan above, assign scores using these scales:
|
| 47 |
+
|
| 48 |
+
**Individual Helpfulness Scores (1-5):**
|
| 49 |
+
- 5: Extremely Helpful - Completely aligned with what the user was asking for
|
| 50 |
+
- 4: Mostly Helpful - Generally useful with minor room for improvement
|
| 51 |
+
- 3: Partially Helpful - Misses the overall goal in some way
|
| 52 |
+
- 2: Borderline Unhelpful - Mostly doesn't capture what the user wanted
|
| 53 |
+
- 1: Not Helpful - Completely missed the essence of the request
|
| 54 |
+
|
| 55 |
+
**Comparative Ranking (1-6):**
|
| 56 |
+
- 1: Response 1 is much better than Response 2
|
| 57 |
+
- 2: Response 1 is better than Response 2
|
| 58 |
+
- 3: Response 1 is slightly better than Response 2
|
| 59 |
+
- 4: Response 2 is slightly better than Response 1
|
| 60 |
+
- 5: Response 2 is better than Response 1
|
| 61 |
+
- 6: Response 2 is much better than Response 1
|
| 62 |
+
|
| 63 |
+
#### Output Format ####
|
| 64 |
+
Analyze step by step following the evaluation plan, then provide your judgment as JSON:
|
| 65 |
+
```json
|
| 66 |
+
{
|
| 67 |
+
"response_1_analysis": "Your detailed analysis of Response 1 based on the evaluation plan",
|
| 68 |
+
"response_2_analysis": "Your detailed analysis of Response 2 based on the evaluation plan",
|
| 69 |
+
"score_1": <1-5>,
|
| 70 |
+
"score_2": <1-5>,
|
| 71 |
+
"ranking": <1-6>
|
| 72 |
+
}
|
| 73 |
+
```
|
| 74 |
+
{{- '<|im_end|>\n' }}
|
| 75 |
+
{%- endif %}
|
| 76 |
+
{%- set ns = namespace(multi_step_tool=true, last_query_index=messages|length - 1) %}
|
| 77 |
+
{%- for message in messages[::-1] %}
|
| 78 |
+
{%- set index = (messages|length - 1) - loop.index0 %}
|
| 79 |
+
{%- if ns.multi_step_tool and message.role == "user" and message.content is string and not(message.content.startswith('<tool_response>') and message.content.endswith('</tool_response>')) %}
|
| 80 |
+
{%- set ns.multi_step_tool = false %}
|
| 81 |
+
{%- set ns.last_query_index = index %}
|
| 82 |
+
{%- endif %}
|
| 83 |
+
{%- endfor %}
|
| 84 |
+
{%- for message in messages %}
|
| 85 |
+
{%- if message.content is string %}
|
| 86 |
+
{%- set content = message.content %}
|
| 87 |
+
{%- else %}
|
| 88 |
+
{%- set content = '' %}
|
| 89 |
+
{%- endif %}
|
| 90 |
+
{%- if (message.role == "user") or (message.role == "system" and not loop.first) %}
|
| 91 |
+
{{- '<|im_start|>' + message.role + '\n' + content + '<|im_end|>' + '\n' }}
|
| 92 |
+
{%- elif message.role == "assistant" %}
|
| 93 |
+
{%- set reasoning_content = '' %}
|
| 94 |
+
{%- if message.reasoning_content is string %}
|
| 95 |
+
{%- set reasoning_content = message.reasoning_content %}
|
| 96 |
+
{%- else %}
|
| 97 |
+
{%- if '</think>' in content %}
|
| 98 |
+
{%- set reasoning_content = content.split('</think>')[0].rstrip('\n').split('<think>')[-1].lstrip('\n') %}
|
| 99 |
+
{%- set content = content.split('</think>')[-1].lstrip('\n') %}
|
| 100 |
+
{%- endif %}
|
| 101 |
+
{%- endif %}
|
| 102 |
+
{%- if loop.index0 > ns.last_query_index %}
|
| 103 |
+
{%- if loop.last or (not loop.last and reasoning_content) %}
|
| 104 |
+
{{- '<|im_start|>' + message.role + '\n<think>\n' + reasoning_content.strip('\n') + '\n</think>\n\n' + content.lstrip('\n') }}
|
| 105 |
+
{%- else %}
|
| 106 |
+
{{- '<|im_start|>' + message.role + '\n' + content }}
|
| 107 |
+
{%- endif %}
|
| 108 |
+
{%- else %}
|
| 109 |
+
{{- '<|im_start|>' + message.role + '\n' + content }}
|
| 110 |
+
{%- endif %}
|
| 111 |
+
{%- if message.tool_calls %}
|
| 112 |
+
{%- for tool_call in message.tool_calls %}
|
| 113 |
+
{%- if (loop.first and content) or (not loop.first) %}
|
| 114 |
+
{{- '\n' }}
|
| 115 |
+
{%- endif %}
|
| 116 |
+
{%- if tool_call.function %}
|
| 117 |
+
{%- set tool_call = tool_call.function %}
|
| 118 |
+
{%- endif %}
|
| 119 |
+
{{- '<tool_call>\n{"name": "' }}
|
| 120 |
+
{{- tool_call.name }}
|
| 121 |
+
{{- '", "arguments": ' }}
|
| 122 |
+
{%- if tool_call.arguments is string %}
|
| 123 |
+
{{- tool_call.arguments }}
|
| 124 |
+
{%- else %}
|
| 125 |
+
{{- tool_call.arguments | tojson }}
|
| 126 |
+
{%- endif %}
|
| 127 |
+
{{- '}\n</tool_call>' }}
|
| 128 |
+
{%- endfor %}
|
| 129 |
+
{%- endif %}
|
| 130 |
+
{{- '<|im_end|>\n' }}
|
| 131 |
+
{%- elif message.role == "tool" %}
|
| 132 |
+
{%- if loop.first or (messages[loop.index0 - 1].role != "tool") %}
|
| 133 |
+
{{- '<|im_start|>user' }}
|
| 134 |
+
{%- endif %}
|
| 135 |
+
{{- '\n<tool_response>\n' }}
|
| 136 |
+
{{- content }}
|
| 137 |
+
{{- '\n</tool_response>' }}
|
| 138 |
+
{%- if loop.last or (messages[loop.index0 + 1].role != "tool") %}
|
| 139 |
+
{{- '<|im_end|>\n' }}
|
| 140 |
+
{%- endif %}
|
| 141 |
+
{%- endif %}
|
| 142 |
+
{%- endfor %}
|
| 143 |
+
{%- if add_generation_prompt %}
|
| 144 |
+
{{- '<|im_start|>assistant\n' }}
|
| 145 |
+
{%- if enable_thinking is defined and enable_thinking is false %}
|
| 146 |
+
{{- '<think>\n\n</think>\n\n' }}
|
| 147 |
+
{%- endif %}
|
| 148 |
+
{%- endif %}
|
config.json
ADDED
|
@@ -0,0 +1,38 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"architectures": [
|
| 3 |
+
"Qwen3MoeForCausalLM"
|
| 4 |
+
],
|
| 5 |
+
"attention_bias": false,
|
| 6 |
+
"attention_dropout": 0.0,
|
| 7 |
+
"bos_token_id": 151643,
|
| 8 |
+
"decoder_sparse_step": 1,
|
| 9 |
+
"eos_token_id": 151645,
|
| 10 |
+
"head_dim": 128,
|
| 11 |
+
"hidden_act": "silu",
|
| 12 |
+
"hidden_size": 4096,
|
| 13 |
+
"initializer_range": 0.02,
|
| 14 |
+
"intermediate_size": 12288,
|
| 15 |
+
"max_position_embeddings": 40960,
|
| 16 |
+
"max_window_layers": 94,
|
| 17 |
+
"mlp_only_layers": [],
|
| 18 |
+
"model_type": "qwen3_moe",
|
| 19 |
+
"moe_intermediate_size": 1536,
|
| 20 |
+
"norm_topk_prob": true,
|
| 21 |
+
"num_attention_heads": 64,
|
| 22 |
+
"num_experts": 128,
|
| 23 |
+
"num_experts_per_tok": 8,
|
| 24 |
+
"num_hidden_layers": 94,
|
| 25 |
+
"num_key_value_heads": 4,
|
| 26 |
+
"output_router_logits": false,
|
| 27 |
+
"rms_norm_eps": 1e-06,
|
| 28 |
+
"rope_scaling": null,
|
| 29 |
+
"rope_theta": 5000000,
|
| 30 |
+
"router_aux_loss_coef": 0.001,
|
| 31 |
+
"sliding_window": null,
|
| 32 |
+
"tie_word_embeddings": false,
|
| 33 |
+
"torch_dtype": "bfloat16",
|
| 34 |
+
"transformers_version": "4.53.3",
|
| 35 |
+
"use_cache": true,
|
| 36 |
+
"use_sliding_window": false,
|
| 37 |
+
"vocab_size": 151936
|
| 38 |
+
}
|
explainability.md
ADDED
|
@@ -0,0 +1,14 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
|
| 2 |
+
| Field | Response |
|
| 3 |
+
| ----- | ----- |
|
| 4 |
+
| Intended Application & Domain: | Generative reward model, suitable for LLM-as-a-Judge and Reinforcement Learning from Human Feedback. |
|
| 5 |
+
| Model Type: | Qwen3-235B-A22B MoE |
|
| 6 |
+
| Intended User: | Developers designing AI Agent systems, chatbots, RAG systems, and other AI-powered applications. |
|
| 7 |
+
| Output: | Text (String, One-Dimensional sequences). |
|
| 8 |
+
| Describe how the model works: | Given a conversation history, a user request, and two candidate responses, the model generates a reasoning trace then an individual helpfulness score for each response and a ranking score. |
|
| 9 |
+
| Name the adversely impacted groups this has been tested to deliver comparable outcomes regardless of: | N/A |
|
| 10 |
+
| Technical Limitations: | The model's max sequence length is 128K tokens. Longer text inputs should be truncated. The model may show decreased accuracy for harder prompts when configured to skip intermediate reasoning traces. |
|
| 11 |
+
| Verified to have met prescribed NVIDIA quality standards: | Yes |
|
| 12 |
+
| Performance Metrics: | Accuracy, Throughput, and Latency. |
|
| 13 |
+
| Potential Known Risks: | The model was trained on data that contains toxic language and societal biases originally crawled from the internet. Therefore, the model may amplify those biases and return toxic responses especially when prompted with toxic prompts. The model may generate answers that may be inaccurate, omit key information, or include irrelevant or redundant text producing socially unacceptable or undesirable text, even if the prompt itself does not include anything explicitly offensive. |
|
| 14 |
+
| Licensing & Terms of Use: | [Apache 2.0](https://www.apache.org/licenses/LICENSE-2.0.txt) |
|
generation_config.json
ADDED
|
@@ -0,0 +1,13 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"bos_token_id": 151643,
|
| 3 |
+
"do_sample": true,
|
| 4 |
+
"eos_token_id": [
|
| 5 |
+
151645,
|
| 6 |
+
151643
|
| 7 |
+
],
|
| 8 |
+
"pad_token_id": 151643,
|
| 9 |
+
"temperature": 0.6,
|
| 10 |
+
"top_k": 20,
|
| 11 |
+
"top_p": 0.95,
|
| 12 |
+
"transformers_version": "4.53.3"
|
| 13 |
+
}
|
merges.txt
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
model-00001-of-00095.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:bebfa6a00983675b8d6db561c27c59bb6327d3efe8e81464143902539d39821a
|
| 3 |
+
size 4999647472
|
model-00002-of-00095.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:ffb108d472b92e3ee1b7ab255946b528546223230a8eaa85be1872a33ad1785f
|
| 3 |
+
size 4988141728
|
model-00003-of-00095.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:688b397ade6a5a13a7886174716aacf2593aaf2afbfd32cdb37a2c567d7effe7
|
| 3 |
+
size 4988141728
|
model-00004-of-00095.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:657781de5e6bc6b04e560ade4a6808e7da7d5ae08552a15b4f57b886e1a3bb20
|
| 3 |
+
size 4988141728
|
model-00005-of-00095.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:ee4e5bdc477e2f02d31365c1eb3020a048d0a1e4900fdd6b3ce981319d9cd60b
|
| 3 |
+
size 4988141728
|
model-00006-of-00095.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:b2b046192097b429a298a5f9212ad5063644a7f614d4779346960da30c8a80ed
|
| 3 |
+
size 4988141728
|
model-00007-of-00095.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:bf2d6d31adcbdb760eae2e23969c70174701baa671846504ac5b014ad3c1602a
|
| 3 |
+
size 4988141728
|
model-00008-of-00095.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:fb3c576af6d12f136ea2230f99a31b62f4ed3437bb336292cc0220179eb37117
|
| 3 |
+
size 4988141728
|
model-00009-of-00095.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:3e62a09d930c6169a07957d5e0889ca8b07ae91734095d2d76333ad51fb82e01
|
| 3 |
+
size 4988141728
|
model-00010-of-00095.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:d6217ef51013249a343addb3bba0222cf4fe1e9c9303a4b853c57ab5616420a8
|
| 3 |
+
size 4988141728
|
model-00011-of-00095.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:6d2857fc46a3c11df57cf9fb6353a8e08c33ea03d40caef004830d0d15cb9a58
|
| 3 |
+
size 4988142048
|
model-00012-of-00095.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:0ddbdd696b59c628100c509a1568bf74413d6b00d03381700c956cfdfc4cb96a
|
| 3 |
+
size 4988142120
|
model-00013-of-00095.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:9ac11f186889a5faa0aea922297aedeca92f3e3a9e177968fe36399308170746
|
| 3 |
+
size 4988142120
|
model-00014-of-00095.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:fc976d447e4c4ea009fd3c63ceceb17ff0addf13e509fdd24dded9bf23ecf9f4
|
| 3 |
+
size 4988142120
|
model-00015-of-00095.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:6c84a4b82c1fb44eeaad9888de86046fd1d0b1e502bea2f554c48525515ea5d6
|
| 3 |
+
size 4988142120
|
model-00016-of-00095.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:9845f8983eef03e6ac39b9282ca14a1fe07b026d3950ead90edfaeece18b38c5
|
| 3 |
+
size 4988142120
|
model-00017-of-00095.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:50738dca121d760bca0d95c6f8f2967dffd2e3e5c8c355740aa9b8e8c9655488
|
| 3 |
+
size 4988142120
|
model-00018-of-00095.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:a95a791a37a91857f87427453f06a73b5b098f2393c48148a7bcf1cc32340607
|
| 3 |
+
size 4988142120
|
model-00019-of-00095.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:d598db7ebcc6694a8b10b06c7844aaae2f2d5d5e5dbdacb6dd3147eb0a17726e
|
| 3 |
+
size 4988142120
|
model-00020-of-00095.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:f069d2eb0a87ae873a695c2843323d6ffba2ba90fd844945bf94ec0ff0eaab54
|
| 3 |
+
size 4988142120
|
model-00021-of-00095.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:6b401cca518051ae40c51643691b47cd5cdb5738ddd355473ca9e5ad2c649daf
|
| 3 |
+
size 4988142120
|
model-00022-of-00095.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:176246f089dcd7e34282939c293be0bb51af9665d919d3229cbfabb54b6cad3e
|
| 3 |
+
size 4988142120
|
model-00023-of-00095.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:0958c5ef85e34e8251517eac3b11158a5a5dfe0b8043e589e1a4808e7cbacfde
|
| 3 |
+
size 4988142120
|
model-00024-of-00095.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:ec195768088884816d03a8ec413cc40d395aff17981778d6772c57bbc26a224f
|
| 3 |
+
size 4988142120
|
model-00025-of-00095.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:0eab6d85dec15579c4b1bc2a5fb9b506bc927dc7a5e85440efa4bd3c125ecf82
|
| 3 |
+
size 4988142120
|
model-00026-of-00095.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:162feb930d70e96a9078de30c31bcbf30d870d51d3a3326b68ebb40aaf07c6dd
|
| 3 |
+
size 4988142120
|
model-00027-of-00095.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:0f46243ad0ad75a96f911d4b90ee3f3d9b1e6d29b1a90b2f110fb97a07b3acac
|
| 3 |
+
size 4988142120
|
model-00028-of-00095.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:f92d8c754f8aa19bd65bdce114b1bbd0b1dc1895a89e8b5366ea9ae399882dac
|
| 3 |
+
size 4988142120
|
model-00029-of-00095.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:b7d0bccc42fc9f7ecc48508a2a100206f5d768118bf39c3a9db3eefc34a9ec42
|
| 3 |
+
size 4988142120
|
model-00030-of-00095.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:0875b3a8e72199983cdd14b187a5d79db2a77aefc6828d87a95844bfbc15328f
|
| 3 |
+
size 4988142120
|
model-00031-of-00095.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:039c9297037365a27e38bd271a6b4c6b7af597b41fc6bcc86271612783338f4c
|
| 3 |
+
size 4988142120
|
model-00032-of-00095.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:b0c535ae77de4fb83152ae936dc41d8e41036639a220ec0e7b8cf4c323eaf194
|
| 3 |
+
size 4988142120
|
model-00033-of-00095.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:41f3509532736aa224b7ee46bf9145476e44f4a07aac1b2fb660d64e4b13ff5c
|
| 3 |
+
size 4988142120
|
model-00034-of-00095.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:71252f3f4d870504d0ce3899d8014035a085a31f154676c101d8e8a7aea53b95
|
| 3 |
+
size 4988142120
|
model-00035-of-00095.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:b47f30ac3bd315fa2650b70c767a81b9f26e8ce63c6562dc60caa12f20b035f2
|
| 3 |
+
size 4988142120
|
model-00036-of-00095.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:24d064856ac9edab5b7d5a018a35f9312b68e8c5d9b78205e9986af16b225291
|
| 3 |
+
size 4988142120
|
model-00037-of-00095.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:604cb4134e8c3475d5a74f969651d86f323ecec6f1624881b2679748e85d9b57
|
| 3 |
+
size 4988142120
|
model-00038-of-00095.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:501a04a90fe296a7b802f50fffadfabd65171e4d5ef8e520bb9f35e500e7969a
|
| 3 |
+
size 4988142120
|
model-00039-of-00095.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:8def596e403ef4c4d14e67cf362643ec5fedda9b9c14fab454dd01743260c1ac
|
| 3 |
+
size 4988142120
|