hai2131 commited on
Commit
e8b5db8
·
verified ·
1 Parent(s): 94271d9

Upload checkpoint with token

Browse files
.gitattributes CHANGED
@@ -33,3 +33,4 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
 
 
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
36
+ tokenizer.json filter=lfs diff=lfs merge=lfs -text
README.md ADDED
@@ -0,0 +1,202 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ base_model: hai2131/sailor2-sft
3
+ library_name: peft
4
+ ---
5
+
6
+ # Model Card for Model ID
7
+
8
+ <!-- Provide a quick summary of what the model is/does. -->
9
+
10
+
11
+
12
+ ## Model Details
13
+
14
+ ### Model Description
15
+
16
+ <!-- Provide a longer summary of what this model is. -->
17
+
18
+
19
+
20
+ - **Developed by:** [More Information Needed]
21
+ - **Funded by [optional]:** [More Information Needed]
22
+ - **Shared by [optional]:** [More Information Needed]
23
+ - **Model type:** [More Information Needed]
24
+ - **Language(s) (NLP):** [More Information Needed]
25
+ - **License:** [More Information Needed]
26
+ - **Finetuned from model [optional]:** [More Information Needed]
27
+
28
+ ### Model Sources [optional]
29
+
30
+ <!-- Provide the basic links for the model. -->
31
+
32
+ - **Repository:** [More Information Needed]
33
+ - **Paper [optional]:** [More Information Needed]
34
+ - **Demo [optional]:** [More Information Needed]
35
+
36
+ ## Uses
37
+
38
+ <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
39
+
40
+ ### Direct Use
41
+
42
+ <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
43
+
44
+ [More Information Needed]
45
+
46
+ ### Downstream Use [optional]
47
+
48
+ <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
49
+
50
+ [More Information Needed]
51
+
52
+ ### Out-of-Scope Use
53
+
54
+ <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
55
+
56
+ [More Information Needed]
57
+
58
+ ## Bias, Risks, and Limitations
59
+
60
+ <!-- This section is meant to convey both technical and sociotechnical limitations. -->
61
+
62
+ [More Information Needed]
63
+
64
+ ### Recommendations
65
+
66
+ <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
67
+
68
+ Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
69
+
70
+ ## How to Get Started with the Model
71
+
72
+ Use the code below to get started with the model.
73
+
74
+ [More Information Needed]
75
+
76
+ ## Training Details
77
+
78
+ ### Training Data
79
+
80
+ <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
81
+
82
+ [More Information Needed]
83
+
84
+ ### Training Procedure
85
+
86
+ <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
87
+
88
+ #### Preprocessing [optional]
89
+
90
+ [More Information Needed]
91
+
92
+
93
+ #### Training Hyperparameters
94
+
95
+ - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
96
+
97
+ #### Speeds, Sizes, Times [optional]
98
+
99
+ <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
100
+
101
+ [More Information Needed]
102
+
103
+ ## Evaluation
104
+
105
+ <!-- This section describes the evaluation protocols and provides the results. -->
106
+
107
+ ### Testing Data, Factors & Metrics
108
+
109
+ #### Testing Data
110
+
111
+ <!-- This should link to a Dataset Card if possible. -->
112
+
113
+ [More Information Needed]
114
+
115
+ #### Factors
116
+
117
+ <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
118
+
119
+ [More Information Needed]
120
+
121
+ #### Metrics
122
+
123
+ <!-- These are the evaluation metrics being used, ideally with a description of why. -->
124
+
125
+ [More Information Needed]
126
+
127
+ ### Results
128
+
129
+ [More Information Needed]
130
+
131
+ #### Summary
132
+
133
+
134
+
135
+ ## Model Examination [optional]
136
+
137
+ <!-- Relevant interpretability work for the model goes here -->
138
+
139
+ [More Information Needed]
140
+
141
+ ## Environmental Impact
142
+
143
+ <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
144
+
145
+ Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
146
+
147
+ - **Hardware Type:** [More Information Needed]
148
+ - **Hours used:** [More Information Needed]
149
+ - **Cloud Provider:** [More Information Needed]
150
+ - **Compute Region:** [More Information Needed]
151
+ - **Carbon Emitted:** [More Information Needed]
152
+
153
+ ## Technical Specifications [optional]
154
+
155
+ ### Model Architecture and Objective
156
+
157
+ [More Information Needed]
158
+
159
+ ### Compute Infrastructure
160
+
161
+ [More Information Needed]
162
+
163
+ #### Hardware
164
+
165
+ [More Information Needed]
166
+
167
+ #### Software
168
+
169
+ [More Information Needed]
170
+
171
+ ## Citation [optional]
172
+
173
+ <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
174
+
175
+ **BibTeX:**
176
+
177
+ [More Information Needed]
178
+
179
+ **APA:**
180
+
181
+ [More Information Needed]
182
+
183
+ ## Glossary [optional]
184
+
185
+ <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
186
+
187
+ [More Information Needed]
188
+
189
+ ## More Information [optional]
190
+
191
+ [More Information Needed]
192
+
193
+ ## Model Card Authors [optional]
194
+
195
+ [More Information Needed]
196
+
197
+ ## Model Card Contact
198
+
199
+ [More Information Needed]
200
+ ### Framework versions
201
+
202
+ - PEFT 0.15.2
adapter_config.json ADDED
@@ -0,0 +1,39 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "alpha_pattern": {},
3
+ "auto_mapping": null,
4
+ "base_model_name_or_path": "hai2131/sailor2-sft",
5
+ "bias": "none",
6
+ "corda_config": null,
7
+ "eva_config": null,
8
+ "exclude_modules": null,
9
+ "fan_in_fan_out": false,
10
+ "inference_mode": true,
11
+ "init_lora_weights": true,
12
+ "layer_replication": null,
13
+ "layers_pattern": null,
14
+ "layers_to_transform": null,
15
+ "loftq_config": {},
16
+ "lora_alpha": 64,
17
+ "lora_bias": false,
18
+ "lora_dropout": 0.0,
19
+ "megatron_config": null,
20
+ "megatron_core": "megatron.core",
21
+ "modules_to_save": null,
22
+ "peft_type": "LORA",
23
+ "r": 32,
24
+ "rank_pattern": {},
25
+ "revision": null,
26
+ "target_modules": [
27
+ "o_proj",
28
+ "gate_proj",
29
+ "k_proj",
30
+ "up_proj",
31
+ "q_proj",
32
+ "down_proj",
33
+ "v_proj"
34
+ ],
35
+ "task_type": "CAUSAL_LM",
36
+ "trainable_token_indices": null,
37
+ "use_dora": false,
38
+ "use_rslora": false
39
+ }
adapter_model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c3d9817cf262c8f4b78c429ef51037a3c417c0177d938d5ff276103587fd6d06
3
+ size 140860656
added_tokens.json ADDED
@@ -0,0 +1,24 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "</tool_call>": 151658,
3
+ "<tool_call>": 151657,
4
+ "<|box_end|>": 151649,
5
+ "<|box_start|>": 151648,
6
+ "<|endoftext|>": 151643,
7
+ "<|file_sep|>": 151664,
8
+ "<|fim_middle|>": 151660,
9
+ "<|fim_pad|>": 151662,
10
+ "<|fim_prefix|>": 151659,
11
+ "<|fim_suffix|>": 151661,
12
+ "<|im_end|>": 151645,
13
+ "<|im_start|>": 151644,
14
+ "<|image_pad|>": 151655,
15
+ "<|object_ref_end|>": 151647,
16
+ "<|object_ref_start|>": 151646,
17
+ "<|quad_end|>": 151651,
18
+ "<|quad_start|>": 151650,
19
+ "<|repo_name|>": 151663,
20
+ "<|video_pad|>": 151656,
21
+ "<|vision_end|>": 151653,
22
+ "<|vision_pad|>": 151654,
23
+ "<|vision_start|>": 151652
24
+ }
chat_template.jinja ADDED
@@ -0,0 +1,6 @@
 
 
 
 
 
 
 
1
+ {% for message in messages %}{% if loop.first and messages[0]['role'] != 'system' %}{{ '<|im_start|>system
2
+ You are an AI assistant named Sailor2, created by Sea AI Lab. As an AI assistant, you can answer questions in English, Chinese, and Southeast Asian languages such as Burmese, Cebuano, Ilocano, Indonesian, Javanese, Khmer, Lao, Malay, Sundanese, Tagalog, Thai, Vietnamese, and Waray. Your responses should be friendly, unbiased, informative, detailed, and faithful.<|im_end|>
3
+ ' }}{% endif %}{{'<|im_start|>' + message['role'] + '
4
+ ' + message['content'] + '<|im_end|>' + '
5
+ '}}{% endfor %}{% if add_generation_prompt %}{{ '<|im_start|>assistant
6
+ ' }}{% endif %}
merges.txt ADDED
The diff for this file is too large to render. See raw diff
 
optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f9efe7dc75bf541479a10348dc64d800a34b4e599e08c662bf7c176b933e0fc2
3
+ size 72272293
rng_state_0.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7cc3307602102ef5b4699c5d0d0877f4df4e9a8c6ed16159e0eb5f7c205b7a82
3
+ size 15365
rng_state_1.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4a564fd29fdaea985856b48f282cabacb0dffa1dd1a79d98c9e6177e1de774cf
3
+ size 15365
rng_state_2.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2c1fbc115b0e5dd3df1a482bef8a39fc8b8fc0f501ef419324e60d082c1133bc
3
+ size 15365
rng_state_3.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:476ddfa1fd95c624fc018bdacac3fd79e42329b309ca4076c70c58da0694461a
3
+ size 15365
scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bc3bfdba43e140aa8e5cd5fc947cfb23559f9daad7f487c0e1a19d5a1110e00e
3
+ size 1465
special_tokens_map.json ADDED
@@ -0,0 +1,31 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "additional_special_tokens": [
3
+ "<|im_start|>",
4
+ "<|im_end|>",
5
+ "<|object_ref_start|>",
6
+ "<|object_ref_end|>",
7
+ "<|box_start|>",
8
+ "<|box_end|>",
9
+ "<|quad_start|>",
10
+ "<|quad_end|>",
11
+ "<|vision_start|>",
12
+ "<|vision_end|>",
13
+ "<|vision_pad|>",
14
+ "<|image_pad|>",
15
+ "<|video_pad|>"
16
+ ],
17
+ "eos_token": {
18
+ "content": "<|im_end|>",
19
+ "lstrip": false,
20
+ "normalized": false,
21
+ "rstrip": false,
22
+ "single_word": false
23
+ },
24
+ "pad_token": {
25
+ "content": "<|endoftext|>",
26
+ "lstrip": false,
27
+ "normalized": false,
28
+ "rstrip": false,
29
+ "single_word": false
30
+ }
31
+ }
tokenizer.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9c5ae00e602b8860cbd784ba82a8aa14e8feecec692e7076590d014d7b7fdafa
3
+ size 11421896
tokenizer_config.json ADDED
@@ -0,0 +1,214 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "add_bos_token": false,
3
+ "add_prefix_space": false,
4
+ "added_tokens_decoder": {
5
+ "151643": {
6
+ "content": "<|endoftext|>",
7
+ "lstrip": false,
8
+ "normalized": false,
9
+ "rstrip": false,
10
+ "single_word": false,
11
+ "special": true
12
+ },
13
+ "151644": {
14
+ "content": "<|im_start|>",
15
+ "lstrip": false,
16
+ "normalized": false,
17
+ "rstrip": false,
18
+ "single_word": false,
19
+ "special": true
20
+ },
21
+ "151645": {
22
+ "content": "<|im_end|>",
23
+ "lstrip": false,
24
+ "normalized": false,
25
+ "rstrip": false,
26
+ "single_word": false,
27
+ "special": true
28
+ },
29
+ "151646": {
30
+ "content": "<|object_ref_start|>",
31
+ "lstrip": false,
32
+ "normalized": false,
33
+ "rstrip": false,
34
+ "single_word": false,
35
+ "special": true
36
+ },
37
+ "151647": {
38
+ "content": "<|object_ref_end|>",
39
+ "lstrip": false,
40
+ "normalized": false,
41
+ "rstrip": false,
42
+ "single_word": false,
43
+ "special": true
44
+ },
45
+ "151648": {
46
+ "content": "<|box_start|>",
47
+ "lstrip": false,
48
+ "normalized": false,
49
+ "rstrip": false,
50
+ "single_word": false,
51
+ "special": true
52
+ },
53
+ "151649": {
54
+ "content": "<|box_end|>",
55
+ "lstrip": false,
56
+ "normalized": false,
57
+ "rstrip": false,
58
+ "single_word": false,
59
+ "special": true
60
+ },
61
+ "151650": {
62
+ "content": "<|quad_start|>",
63
+ "lstrip": false,
64
+ "normalized": false,
65
+ "rstrip": false,
66
+ "single_word": false,
67
+ "special": true
68
+ },
69
+ "151651": {
70
+ "content": "<|quad_end|>",
71
+ "lstrip": false,
72
+ "normalized": false,
73
+ "rstrip": false,
74
+ "single_word": false,
75
+ "special": true
76
+ },
77
+ "151652": {
78
+ "content": "<|vision_start|>",
79
+ "lstrip": false,
80
+ "normalized": false,
81
+ "rstrip": false,
82
+ "single_word": false,
83
+ "special": true
84
+ },
85
+ "151653": {
86
+ "content": "<|vision_end|>",
87
+ "lstrip": false,
88
+ "normalized": false,
89
+ "rstrip": false,
90
+ "single_word": false,
91
+ "special": true
92
+ },
93
+ "151654": {
94
+ "content": "<|vision_pad|>",
95
+ "lstrip": false,
96
+ "normalized": false,
97
+ "rstrip": false,
98
+ "single_word": false,
99
+ "special": true
100
+ },
101
+ "151655": {
102
+ "content": "<|image_pad|>",
103
+ "lstrip": false,
104
+ "normalized": false,
105
+ "rstrip": false,
106
+ "single_word": false,
107
+ "special": true
108
+ },
109
+ "151656": {
110
+ "content": "<|video_pad|>",
111
+ "lstrip": false,
112
+ "normalized": false,
113
+ "rstrip": false,
114
+ "single_word": false,
115
+ "special": true
116
+ },
117
+ "151657": {
118
+ "content": "<tool_call>",
119
+ "lstrip": false,
120
+ "normalized": false,
121
+ "rstrip": false,
122
+ "single_word": false,
123
+ "special": false
124
+ },
125
+ "151658": {
126
+ "content": "</tool_call>",
127
+ "lstrip": false,
128
+ "normalized": false,
129
+ "rstrip": false,
130
+ "single_word": false,
131
+ "special": false
132
+ },
133
+ "151659": {
134
+ "content": "<|fim_prefix|>",
135
+ "lstrip": false,
136
+ "normalized": false,
137
+ "rstrip": false,
138
+ "single_word": false,
139
+ "special": false
140
+ },
141
+ "151660": {
142
+ "content": "<|fim_middle|>",
143
+ "lstrip": false,
144
+ "normalized": false,
145
+ "rstrip": false,
146
+ "single_word": false,
147
+ "special": false
148
+ },
149
+ "151661": {
150
+ "content": "<|fim_suffix|>",
151
+ "lstrip": false,
152
+ "normalized": false,
153
+ "rstrip": false,
154
+ "single_word": false,
155
+ "special": false
156
+ },
157
+ "151662": {
158
+ "content": "<|fim_pad|>",
159
+ "lstrip": false,
160
+ "normalized": false,
161
+ "rstrip": false,
162
+ "single_word": false,
163
+ "special": false
164
+ },
165
+ "151663": {
166
+ "content": "<|repo_name|>",
167
+ "lstrip": false,
168
+ "normalized": false,
169
+ "rstrip": false,
170
+ "single_word": false,
171
+ "special": false
172
+ },
173
+ "151664": {
174
+ "content": "<|file_sep|>",
175
+ "lstrip": false,
176
+ "normalized": false,
177
+ "rstrip": false,
178
+ "single_word": false,
179
+ "special": false
180
+ }
181
+ },
182
+ "additional_special_tokens": [
183
+ "<|im_start|>",
184
+ "<|im_end|>",
185
+ "<|object_ref_start|>",
186
+ "<|object_ref_end|>",
187
+ "<|box_start|>",
188
+ "<|box_end|>",
189
+ "<|quad_start|>",
190
+ "<|quad_end|>",
191
+ "<|vision_start|>",
192
+ "<|vision_end|>",
193
+ "<|vision_pad|>",
194
+ "<|image_pad|>",
195
+ "<|video_pad|>"
196
+ ],
197
+ "bos_token": null,
198
+ "clean_up_tokenization_spaces": false,
199
+ "eos_token": "<|im_end|>",
200
+ "errors": "replace",
201
+ "extra_special_tokens": {},
202
+ "max_length": 3848,
203
+ "model_max_length": 4096,
204
+ "pad_to_multiple_of": null,
205
+ "pad_token": "<|endoftext|>",
206
+ "pad_token_type_id": 0,
207
+ "padding_side": "right",
208
+ "split_special_tokens": false,
209
+ "stride": 0,
210
+ "tokenizer_class": "Qwen2Tokenizer",
211
+ "truncation_side": "right",
212
+ "truncation_strategy": "longest_first",
213
+ "unk_token": null
214
+ }
trainer_state.json ADDED
@@ -0,0 +1,1234 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_global_step": null,
3
+ "best_metric": null,
4
+ "best_model_checkpoint": null,
5
+ "epoch": 3.0,
6
+ "eval_steps": 500,
7
+ "global_step": 3750,
8
+ "is_hyper_param_search": false,
9
+ "is_local_process_zero": true,
10
+ "is_world_process_zero": true,
11
+ "log_history": [
12
+ {
13
+ "epoch": 0.04,
14
+ "grad_norm": 39.1501579284668,
15
+ "learning_rate": 1.3066666666666667e-06,
16
+ "logits/chosen": 2.069579839706421,
17
+ "logits/rejected": 2.1242542266845703,
18
+ "logps/chosen": -73.9653091430664,
19
+ "logps/rejected": -88.28418731689453,
20
+ "loss": 4.8463,
21
+ "nll_loss": 0.46437376737594604,
22
+ "rewards/accuracies": 0.7049999833106995,
23
+ "rewards/chosen": -7.396530628204346,
24
+ "rewards/margins": 1.431888461112976,
25
+ "rewards/rejected": -8.828418731689453,
26
+ "step": 50
27
+ },
28
+ {
29
+ "epoch": 0.08,
30
+ "grad_norm": 23.26362419128418,
31
+ "learning_rate": 2.64e-06,
32
+ "logits/chosen": 2.039077043533325,
33
+ "logits/rejected": 2.1060538291931152,
34
+ "logps/chosen": -57.755859375,
35
+ "logps/rejected": -72.76531982421875,
36
+ "loss": 3.8951,
37
+ "nll_loss": 0.35950180888175964,
38
+ "rewards/accuracies": 0.7112500071525574,
39
+ "rewards/chosen": -5.77558708190918,
40
+ "rewards/margins": 1.5009456872940063,
41
+ "rewards/rejected": -7.276532173156738,
42
+ "step": 100
43
+ },
44
+ {
45
+ "epoch": 0.12,
46
+ "grad_norm": 29.71898078918457,
47
+ "learning_rate": 3.973333333333333e-06,
48
+ "logits/chosen": 2.017021894454956,
49
+ "logits/rejected": 2.063059091567993,
50
+ "logps/chosen": -52.110809326171875,
51
+ "logps/rejected": -63.39885711669922,
52
+ "loss": 3.5484,
53
+ "nll_loss": 0.3122263252735138,
54
+ "rewards/accuracies": 0.6600000262260437,
55
+ "rewards/chosen": -5.211080551147461,
56
+ "rewards/margins": 1.1288050413131714,
57
+ "rewards/rejected": -6.339885711669922,
58
+ "step": 150
59
+ },
60
+ {
61
+ "epoch": 0.16,
62
+ "grad_norm": 21.129802703857422,
63
+ "learning_rate": 5.306666666666667e-06,
64
+ "logits/chosen": 2.0352795124053955,
65
+ "logits/rejected": 2.09993052482605,
66
+ "logps/chosen": -46.72736740112305,
67
+ "logps/rejected": -59.81150436401367,
68
+ "loss": 3.1762,
69
+ "nll_loss": 0.2814980447292328,
70
+ "rewards/accuracies": 0.7087500095367432,
71
+ "rewards/chosen": -4.672736644744873,
72
+ "rewards/margins": 1.308414340019226,
73
+ "rewards/rejected": -5.9811506271362305,
74
+ "step": 200
75
+ },
76
+ {
77
+ "epoch": 0.2,
78
+ "grad_norm": 25.128114700317383,
79
+ "learning_rate": 6.640000000000001e-06,
80
+ "logits/chosen": 2.0161619186401367,
81
+ "logits/rejected": 2.080204725265503,
82
+ "logps/chosen": -46.92048263549805,
83
+ "logps/rejected": -59.91488265991211,
84
+ "loss": 3.0645,
85
+ "nll_loss": 0.2713809609413147,
86
+ "rewards/accuracies": 0.71875,
87
+ "rewards/chosen": -4.6920485496521,
88
+ "rewards/margins": 1.2994396686553955,
89
+ "rewards/rejected": -5.991488456726074,
90
+ "step": 250
91
+ },
92
+ {
93
+ "epoch": 0.24,
94
+ "grad_norm": 22.990211486816406,
95
+ "learning_rate": 7.973333333333334e-06,
96
+ "logits/chosen": 2.006951332092285,
97
+ "logits/rejected": 2.0681934356689453,
98
+ "logps/chosen": -43.61228942871094,
99
+ "logps/rejected": -58.394710540771484,
100
+ "loss": 2.8512,
101
+ "nll_loss": 0.2565268874168396,
102
+ "rewards/accuracies": 0.7512500286102295,
103
+ "rewards/chosen": -4.361229419708252,
104
+ "rewards/margins": 1.4782426357269287,
105
+ "rewards/rejected": -5.839471817016602,
106
+ "step": 300
107
+ },
108
+ {
109
+ "epoch": 0.28,
110
+ "grad_norm": 17.802993774414062,
111
+ "learning_rate": 9.306666666666667e-06,
112
+ "logits/chosen": 1.9981739521026611,
113
+ "logits/rejected": 2.056286334991455,
114
+ "logps/chosen": -42.770790100097656,
115
+ "logps/rejected": -53.638092041015625,
116
+ "loss": 3.0075,
117
+ "nll_loss": 0.25887373089790344,
118
+ "rewards/accuracies": 0.6875,
119
+ "rewards/chosen": -4.2770795822143555,
120
+ "rewards/margins": 1.0867302417755127,
121
+ "rewards/rejected": -5.363808631896973,
122
+ "step": 350
123
+ },
124
+ {
125
+ "epoch": 0.32,
126
+ "grad_norm": 21.133516311645508,
127
+ "learning_rate": 9.998752338940612e-06,
128
+ "logits/chosen": 1.876654028892517,
129
+ "logits/rejected": 1.9367033243179321,
130
+ "logps/chosen": -38.542076110839844,
131
+ "logps/rejected": -48.9306755065918,
132
+ "loss": 2.6798,
133
+ "nll_loss": 0.22952309250831604,
134
+ "rewards/accuracies": 0.6937500238418579,
135
+ "rewards/chosen": -3.8542075157165527,
136
+ "rewards/margins": 1.0388593673706055,
137
+ "rewards/rejected": -4.893066883087158,
138
+ "step": 400
139
+ },
140
+ {
141
+ "epoch": 0.36,
142
+ "grad_norm": 23.868408203125,
143
+ "learning_rate": 9.988142751731797e-06,
144
+ "logits/chosen": 1.7912520170211792,
145
+ "logits/rejected": 1.8582316637039185,
146
+ "logps/chosen": -38.205177307128906,
147
+ "logps/rejected": -51.677650451660156,
148
+ "loss": 2.6625,
149
+ "nll_loss": 0.23552022874355316,
150
+ "rewards/accuracies": 0.7412499785423279,
151
+ "rewards/chosen": -3.8205177783966064,
152
+ "rewards/margins": 1.3472480773925781,
153
+ "rewards/rejected": -5.167766094207764,
154
+ "step": 450
155
+ },
156
+ {
157
+ "epoch": 0.4,
158
+ "grad_norm": 19.167991638183594,
159
+ "learning_rate": 9.966729958067638e-06,
160
+ "logits/chosen": 1.7662477493286133,
161
+ "logits/rejected": 1.8194670677185059,
162
+ "logps/chosen": -38.83460998535156,
163
+ "logps/rejected": -50.7568359375,
164
+ "loss": 2.7149,
165
+ "nll_loss": 0.23406948149204254,
166
+ "rewards/accuracies": 0.7137500047683716,
167
+ "rewards/chosen": -3.8834612369537354,
168
+ "rewards/margins": 1.1922229528427124,
169
+ "rewards/rejected": -5.075684070587158,
170
+ "step": 500
171
+ },
172
+ {
173
+ "epoch": 0.44,
174
+ "grad_norm": 22.746564865112305,
175
+ "learning_rate": 9.934560333291077e-06,
176
+ "logits/chosen": 1.6905887126922607,
177
+ "logits/rejected": 1.756126046180725,
178
+ "logps/chosen": -38.07200241088867,
179
+ "logps/rejected": -50.685115814208984,
180
+ "loss": 2.6286,
181
+ "nll_loss": 0.22550682723522186,
182
+ "rewards/accuracies": 0.7250000238418579,
183
+ "rewards/chosen": -3.8072004318237305,
184
+ "rewards/margins": 1.2613120079040527,
185
+ "rewards/rejected": -5.068512439727783,
186
+ "step": 550
187
+ },
188
+ {
189
+ "epoch": 0.48,
190
+ "grad_norm": 19.4133358001709,
191
+ "learning_rate": 9.89170354964594e-06,
192
+ "logits/chosen": 1.7051059007644653,
193
+ "logits/rejected": 1.7545236349105835,
194
+ "logps/chosen": -35.82815933227539,
195
+ "logps/rejected": -47.575111389160156,
196
+ "loss": 2.5393,
197
+ "nll_loss": 0.21860088407993317,
198
+ "rewards/accuracies": 0.7112500071525574,
199
+ "rewards/chosen": -3.5828158855438232,
200
+ "rewards/margins": 1.174695611000061,
201
+ "rewards/rejected": -4.757511138916016,
202
+ "step": 600
203
+ },
204
+ {
205
+ "epoch": 0.52,
206
+ "grad_norm": 16.378013610839844,
207
+ "learning_rate": 9.83825242538238e-06,
208
+ "logits/chosen": 1.6559092998504639,
209
+ "logits/rejected": 1.7158703804016113,
210
+ "logps/chosen": -36.47942352294922,
211
+ "logps/rejected": -49.875587463378906,
212
+ "loss": 2.5252,
213
+ "nll_loss": 0.21864959597587585,
214
+ "rewards/accuracies": 0.7337499856948853,
215
+ "rewards/chosen": -3.647942304611206,
216
+ "rewards/margins": 1.3396165370941162,
217
+ "rewards/rejected": -4.987558364868164,
218
+ "step": 650
219
+ },
220
+ {
221
+ "epoch": 0.56,
222
+ "grad_norm": 15.000406265258789,
223
+ "learning_rate": 9.774322723733216e-06,
224
+ "logits/chosen": 1.6370524168014526,
225
+ "logits/rejected": 1.6942112445831299,
226
+ "logps/chosen": -34.71821212768555,
227
+ "logps/rejected": -48.40507507324219,
228
+ "loss": 2.4278,
229
+ "nll_loss": 0.21341566741466522,
230
+ "rewards/accuracies": 0.7549999952316284,
231
+ "rewards/chosen": -3.471820592880249,
232
+ "rewards/margins": 1.368686556816101,
233
+ "rewards/rejected": -4.8405070304870605,
234
+ "step": 700
235
+ },
236
+ {
237
+ "epoch": 0.6,
238
+ "grad_norm": 17.42306900024414,
239
+ "learning_rate": 9.700052902196541e-06,
240
+ "logits/chosen": 1.606693148612976,
241
+ "logits/rejected": 1.6712034940719604,
242
+ "logps/chosen": -34.60812759399414,
243
+ "logps/rejected": -48.792850494384766,
244
+ "loss": 2.4102,
245
+ "nll_loss": 0.21256333589553833,
246
+ "rewards/accuracies": 0.768750011920929,
247
+ "rewards/chosen": -3.460812568664551,
248
+ "rewards/margins": 1.4184722900390625,
249
+ "rewards/rejected": -4.879284858703613,
250
+ "step": 750
251
+ },
252
+ {
253
+ "epoch": 0.64,
254
+ "grad_norm": 18.26099967956543,
255
+ "learning_rate": 9.615603812667618e-06,
256
+ "logits/chosen": 1.574639916419983,
257
+ "logits/rejected": 1.6343886852264404,
258
+ "logps/chosen": -34.159088134765625,
259
+ "logps/rejected": -48.842376708984375,
260
+ "loss": 2.3734,
261
+ "nll_loss": 0.20658308267593384,
262
+ "rewards/accuracies": 0.7512500286102295,
263
+ "rewards/chosen": -3.4159088134765625,
264
+ "rewards/margins": 1.4683291912078857,
265
+ "rewards/rejected": -4.884237766265869,
266
+ "step": 800
267
+ },
268
+ {
269
+ "epoch": 0.68,
270
+ "grad_norm": 19.284168243408203,
271
+ "learning_rate": 9.521158353069494e-06,
272
+ "logits/chosen": 1.5315083265304565,
273
+ "logits/rejected": 1.5860241651535034,
274
+ "logps/chosen": -35.90006637573242,
275
+ "logps/rejected": -51.2423095703125,
276
+ "loss": 2.4185,
277
+ "nll_loss": 0.2139773815870285,
278
+ "rewards/accuracies": 0.7487499713897705,
279
+ "rewards/chosen": -3.5900063514709473,
280
+ "rewards/margins": 1.5342237949371338,
281
+ "rewards/rejected": -5.124230861663818,
282
+ "step": 850
283
+ },
284
+ {
285
+ "epoch": 0.72,
286
+ "grad_norm": 28.010725021362305,
287
+ "learning_rate": 9.416921071236821e-06,
288
+ "logits/chosen": 1.562249779701233,
289
+ "logits/rejected": 1.6213380098342896,
290
+ "logps/chosen": -34.88312530517578,
291
+ "logps/rejected": -49.12727355957031,
292
+ "loss": 2.4402,
293
+ "nll_loss": 0.20900359749794006,
294
+ "rewards/accuracies": 0.7362499833106995,
295
+ "rewards/chosen": -3.4883129596710205,
296
+ "rewards/margins": 1.4244145154953003,
297
+ "rewards/rejected": -4.912726879119873,
298
+ "step": 900
299
+ },
300
+ {
301
+ "epoch": 0.76,
302
+ "grad_norm": 22.14881134033203,
303
+ "learning_rate": 9.303117721910801e-06,
304
+ "logits/chosen": 1.5611748695373535,
305
+ "logits/rejected": 1.613703966140747,
306
+ "logps/chosen": -33.08067321777344,
307
+ "logps/rejected": -46.956138610839844,
308
+ "loss": 2.2957,
309
+ "nll_loss": 0.19786672294139862,
310
+ "rewards/accuracies": 0.7524999976158142,
311
+ "rewards/chosen": -3.3080673217773438,
312
+ "rewards/margins": 1.3875468969345093,
313
+ "rewards/rejected": -4.695613861083984,
314
+ "step": 950
315
+ },
316
+ {
317
+ "epoch": 0.8,
318
+ "grad_norm": 18.468719482421875,
319
+ "learning_rate": 9.179994777804677e-06,
320
+ "logits/chosen": 1.532340407371521,
321
+ "logits/rejected": 1.595700979232788,
322
+ "logps/chosen": -33.659690856933594,
323
+ "logps/rejected": -48.1495475769043,
324
+ "loss": 2.3428,
325
+ "nll_loss": 0.2023465782403946,
326
+ "rewards/accuracies": 0.7562500238418579,
327
+ "rewards/chosen": -3.365969657897949,
328
+ "rewards/margins": 1.4489855766296387,
329
+ "rewards/rejected": -4.81495475769043,
330
+ "step": 1000
331
+ },
332
+ {
333
+ "epoch": 0.84,
334
+ "grad_norm": 19.125810623168945,
335
+ "learning_rate": 9.04781889579873e-06,
336
+ "logits/chosen": 1.4953616857528687,
337
+ "logits/rejected": 1.5516945123672485,
338
+ "logps/chosen": -33.457584381103516,
339
+ "logps/rejected": -48.823360443115234,
340
+ "loss": 2.2845,
341
+ "nll_loss": 0.2007952332496643,
342
+ "rewards/accuracies": 0.7612500190734863,
343
+ "rewards/chosen": -3.3457586765289307,
344
+ "rewards/margins": 1.5365774631500244,
345
+ "rewards/rejected": -4.882336139678955,
346
+ "step": 1050
347
+ },
348
+ {
349
+ "epoch": 0.88,
350
+ "grad_norm": 18.376596450805664,
351
+ "learning_rate": 8.906876339420851e-06,
352
+ "logits/chosen": 1.452122449874878,
353
+ "logits/rejected": 1.519229531288147,
354
+ "logps/chosen": -31.664844512939453,
355
+ "logps/rejected": -47.539306640625,
356
+ "loss": 2.1647,
357
+ "nll_loss": 0.19097968935966492,
358
+ "rewards/accuracies": 0.7925000190734863,
359
+ "rewards/chosen": -3.1664845943450928,
360
+ "rewards/margins": 1.5874464511871338,
361
+ "rewards/rejected": -4.753930568695068,
362
+ "step": 1100
363
+ },
364
+ {
365
+ "epoch": 0.92,
366
+ "grad_norm": 20.535789489746094,
367
+ "learning_rate": 8.757472358863481e-06,
368
+ "logits/chosen": 1.4528400897979736,
369
+ "logits/rejected": 1.5169428586959839,
370
+ "logps/chosen": -33.37004089355469,
371
+ "logps/rejected": -47.855865478515625,
372
+ "loss": 2.2511,
373
+ "nll_loss": 0.19547508656978607,
374
+ "rewards/accuracies": 0.7574999928474426,
375
+ "rewards/chosen": -3.337003707885742,
376
+ "rewards/margins": 1.4485827684402466,
377
+ "rewards/rejected": -4.785586357116699,
378
+ "step": 1150
379
+ },
380
+ {
381
+ "epoch": 0.96,
382
+ "grad_norm": 22.283950805664062,
383
+ "learning_rate": 8.599930529879669e-06,
384
+ "logits/chosen": 1.4502493143081665,
385
+ "logits/rejected": 1.5124887228012085,
386
+ "logps/chosen": -31.987987518310547,
387
+ "logps/rejected": -47.01323699951172,
388
+ "loss": 2.1816,
389
+ "nll_loss": 0.19316434860229492,
390
+ "rewards/accuracies": 0.7774999737739563,
391
+ "rewards/chosen": -3.198798418045044,
392
+ "rewards/margins": 1.5025256872177124,
393
+ "rewards/rejected": -4.701324462890625,
394
+ "step": 1200
395
+ },
396
+ {
397
+ "epoch": 1.0,
398
+ "grad_norm": 19.24060821533203,
399
+ "learning_rate": 8.434592052990044e-06,
400
+ "logits/chosen": 1.4031856060028076,
401
+ "logits/rejected": 1.462039589881897,
402
+ "logps/chosen": -30.157163619995117,
403
+ "logps/rejected": -45.0657844543457,
404
+ "loss": 2.1647,
405
+ "nll_loss": 0.18531297147274017,
406
+ "rewards/accuracies": 0.7662500143051147,
407
+ "rewards/chosen": -3.015716314315796,
408
+ "rewards/margins": 1.4908628463745117,
409
+ "rewards/rejected": -4.5065789222717285,
410
+ "step": 1250
411
+ },
412
+ {
413
+ "epoch": 1.04,
414
+ "grad_norm": 16.792695999145508,
415
+ "learning_rate": 8.261815014518465e-06,
416
+ "logits/chosen": 1.334591031074524,
417
+ "logits/rejected": 1.3955891132354736,
418
+ "logps/chosen": -29.710214614868164,
419
+ "logps/rejected": -49.487152099609375,
420
+ "loss": 1.9604,
421
+ "nll_loss": 0.17893625795841217,
422
+ "rewards/accuracies": 0.8450000286102295,
423
+ "rewards/chosen": -2.9710214138031006,
424
+ "rewards/margins": 1.977694034576416,
425
+ "rewards/rejected": -4.948715686798096,
426
+ "step": 1300
427
+ },
428
+ {
429
+ "epoch": 1.08,
430
+ "grad_norm": 21.147741317749023,
431
+ "learning_rate": 8.081973611056784e-06,
432
+ "logits/chosen": 1.257203221321106,
433
+ "logits/rejected": 1.3112421035766602,
434
+ "logps/chosen": -28.43024444580078,
435
+ "logps/rejected": -48.68648910522461,
436
+ "loss": 1.872,
437
+ "nll_loss": 0.16993314027786255,
438
+ "rewards/accuracies": 0.8662499785423279,
439
+ "rewards/chosen": -2.843024492263794,
440
+ "rewards/margins": 2.0256245136260986,
441
+ "rewards/rejected": -4.868649482727051,
442
+ "step": 1350
443
+ },
444
+ {
445
+ "epoch": 1.12,
446
+ "grad_norm": 18.095643997192383,
447
+ "learning_rate": 7.89545733903834e-06,
448
+ "logits/chosen": 1.268505573272705,
449
+ "logits/rejected": 1.3283963203430176,
450
+ "logps/chosen": -29.838397979736328,
451
+ "logps/rejected": -48.807472229003906,
452
+ "loss": 1.9653,
453
+ "nll_loss": 0.18002687394618988,
454
+ "rewards/accuracies": 0.8500000238418579,
455
+ "rewards/chosen": -2.983839988708496,
456
+ "rewards/margins": 1.8969072103500366,
457
+ "rewards/rejected": -4.880747318267822,
458
+ "step": 1400
459
+ },
460
+ {
461
+ "epoch": 1.16,
462
+ "grad_norm": 20.254247665405273,
463
+ "learning_rate": 7.702670151175435e-06,
464
+ "logits/chosen": 1.3044127225875854,
465
+ "logits/rejected": 1.368518352508545,
466
+ "logps/chosen": -29.120044708251953,
467
+ "logps/rejected": -47.06825637817383,
468
+ "loss": 1.9557,
469
+ "nll_loss": 0.1742786169052124,
470
+ "rewards/accuracies": 0.8387500047683716,
471
+ "rewards/chosen": -2.9120047092437744,
472
+ "rewards/margins": 1.7948211431503296,
473
+ "rewards/rejected": -4.706826210021973,
474
+ "step": 1450
475
+ },
476
+ {
477
+ "epoch": 1.2,
478
+ "grad_norm": 15.50015640258789,
479
+ "learning_rate": 7.5040295815877e-06,
480
+ "logits/chosen": 1.2754545211791992,
481
+ "logits/rejected": 1.33790922164917,
482
+ "logps/chosen": -28.515518188476562,
483
+ "logps/rejected": -49.134769439697266,
484
+ "loss": 1.8568,
485
+ "nll_loss": 0.16948945820331573,
486
+ "rewards/accuracies": 0.8737499713897705,
487
+ "rewards/chosen": -2.8515517711639404,
488
+ "rewards/margins": 2.061924934387207,
489
+ "rewards/rejected": -4.913476943969727,
490
+ "step": 1500
491
+ },
492
+ {
493
+ "epoch": 1.24,
494
+ "grad_norm": 18.263025283813477,
495
+ "learning_rate": 7.299965841516164e-06,
496
+ "logits/chosen": 1.2721493244171143,
497
+ "logits/rejected": 1.3371402025222778,
498
+ "logps/chosen": -29.20949935913086,
499
+ "logps/rejected": -48.70906066894531,
500
+ "loss": 1.9402,
501
+ "nll_loss": 0.17609195411205292,
502
+ "rewards/accuracies": 0.8550000190734863,
503
+ "rewards/chosen": -2.920949935913086,
504
+ "rewards/margins": 1.9499554634094238,
505
+ "rewards/rejected": -4.87090539932251,
506
+ "step": 1550
507
+ },
508
+ {
509
+ "epoch": 1.28,
510
+ "grad_norm": 17.24570655822754,
511
+ "learning_rate": 7.090920887581507e-06,
512
+ "logits/chosen": 1.2389811277389526,
513
+ "logits/rejected": 1.2887934446334839,
514
+ "logps/chosen": -28.364063262939453,
515
+ "logps/rejected": -46.97511291503906,
516
+ "loss": 1.8952,
517
+ "nll_loss": 0.1697738617658615,
518
+ "rewards/accuracies": 0.8362500071525574,
519
+ "rewards/chosen": -2.8364062309265137,
520
+ "rewards/margins": 1.8611053228378296,
521
+ "rewards/rejected": -4.697511196136475,
522
+ "step": 1600
523
+ },
524
+ {
525
+ "epoch": 1.32,
526
+ "grad_norm": 20.073883056640625,
527
+ "learning_rate": 6.877347464604446e-06,
528
+ "logits/chosen": 1.2145293951034546,
529
+ "logits/rejected": 1.2653400897979736,
530
+ "logps/chosen": -28.30023765563965,
531
+ "logps/rejected": -47.83637619018555,
532
+ "loss": 1.8857,
533
+ "nll_loss": 0.16985949873924255,
534
+ "rewards/accuracies": 0.8374999761581421,
535
+ "rewards/chosen": -2.830024003982544,
536
+ "rewards/margins": 1.9536142349243164,
537
+ "rewards/rejected": -4.7836384773254395,
538
+ "step": 1650
539
+ },
540
+ {
541
+ "epoch": 1.3599999999999999,
542
+ "grad_norm": 20.44426727294922,
543
+ "learning_rate": 6.659708125061242e-06,
544
+ "logits/chosen": 1.210806131362915,
545
+ "logits/rejected": 1.2653027772903442,
546
+ "logps/chosen": -28.112794876098633,
547
+ "logps/rejected": -47.89250183105469,
548
+ "loss": 1.8333,
549
+ "nll_loss": 0.16796012222766876,
550
+ "rewards/accuracies": 0.8637499809265137,
551
+ "rewards/chosen": -2.811279535293579,
552
+ "rewards/margins": 1.9779702425003052,
553
+ "rewards/rejected": -4.789250373840332,
554
+ "step": 1700
555
+ },
556
+ {
557
+ "epoch": 1.4,
558
+ "grad_norm": 15.871622085571289,
559
+ "learning_rate": 6.438474227298065e-06,
560
+ "logits/chosen": 1.2420759201049805,
561
+ "logits/rejected": 1.296012282371521,
562
+ "logps/chosen": -28.223691940307617,
563
+ "logps/rejected": -47.52497863769531,
564
+ "loss": 1.8911,
565
+ "nll_loss": 0.1713765263557434,
566
+ "rewards/accuracies": 0.8500000238418579,
567
+ "rewards/chosen": -2.82236909866333,
568
+ "rewards/margins": 1.9301284551620483,
569
+ "rewards/rejected": -4.752498149871826,
570
+ "step": 1750
571
+ },
572
+ {
573
+ "epoch": 1.44,
574
+ "grad_norm": 17.20113182067871,
575
+ "learning_rate": 6.2141249146737545e-06,
576
+ "logits/chosen": 1.2283880710601807,
577
+ "logits/rejected": 1.2828854322433472,
578
+ "logps/chosen": -26.78079605102539,
579
+ "logps/rejected": -45.25880813598633,
580
+ "loss": 1.8039,
581
+ "nll_loss": 0.15980492532253265,
582
+ "rewards/accuracies": 0.8450000286102295,
583
+ "rewards/chosen": -2.678079605102539,
584
+ "rewards/margins": 1.8478012084960938,
585
+ "rewards/rejected": -4.525880813598633,
586
+ "step": 1800
587
+ },
588
+ {
589
+ "epoch": 1.48,
590
+ "grad_norm": 20.9070987701416,
591
+ "learning_rate": 5.987146077842015e-06,
592
+ "logits/chosen": 1.2112573385238647,
593
+ "logits/rejected": 1.2712033987045288,
594
+ "logps/chosen": -28.28427505493164,
595
+ "logps/rejected": -48.554771423339844,
596
+ "loss": 1.8766,
597
+ "nll_loss": 0.1684311032295227,
598
+ "rewards/accuracies": 0.8387500047683716,
599
+ "rewards/chosen": -2.828427314758301,
600
+ "rewards/margins": 2.0270493030548096,
601
+ "rewards/rejected": -4.855477333068848,
602
+ "step": 1850
603
+ },
604
+ {
605
+ "epoch": 1.52,
606
+ "grad_norm": 16.220539093017578,
607
+ "learning_rate": 5.7580293024204455e-06,
608
+ "logits/chosen": 1.2373614311218262,
609
+ "logits/rejected": 1.2950869798660278,
610
+ "logps/chosen": -27.206884384155273,
611
+ "logps/rejected": -46.864036560058594,
612
+ "loss": 1.7983,
613
+ "nll_loss": 0.16223596036434174,
614
+ "rewards/accuracies": 0.8712499737739563,
615
+ "rewards/chosen": -2.720688819885254,
616
+ "rewards/margins": 1.9657151699066162,
617
+ "rewards/rejected": -4.686403751373291,
618
+ "step": 1900
619
+ },
620
+ {
621
+ "epoch": 1.56,
622
+ "grad_norm": 19.87531852722168,
623
+ "learning_rate": 5.5272708043255605e-06,
624
+ "logits/chosen": 1.2204946279525757,
625
+ "logits/rejected": 1.266890048980713,
626
+ "logps/chosen": -26.85431671142578,
627
+ "logps/rejected": -45.090606689453125,
628
+ "loss": 1.8634,
629
+ "nll_loss": 0.1644485890865326,
630
+ "rewards/accuracies": 0.8387500047683716,
631
+ "rewards/chosen": -2.685431718826294,
632
+ "rewards/margins": 1.8236281871795654,
633
+ "rewards/rejected": -4.509060382843018,
634
+ "step": 1950
635
+ },
636
+ {
637
+ "epoch": 1.6,
638
+ "grad_norm": 20.69000244140625,
639
+ "learning_rate": 5.295370355079615e-06,
640
+ "logits/chosen": 1.1966464519500732,
641
+ "logits/rejected": 1.2531131505966187,
642
+ "logps/chosen": -26.715871810913086,
643
+ "logps/rejected": -46.64927673339844,
644
+ "loss": 1.7931,
645
+ "nll_loss": 0.16148775815963745,
646
+ "rewards/accuracies": 0.8424999713897705,
647
+ "rewards/chosen": -2.6715869903564453,
648
+ "rewards/margins": 1.993341088294983,
649
+ "rewards/rejected": -4.664927959442139,
650
+ "step": 2000
651
+ },
652
+ {
653
+ "epoch": 1.6400000000000001,
654
+ "grad_norm": 20.352148056030273,
655
+ "learning_rate": 5.062830199416764e-06,
656
+ "logits/chosen": 1.2095577716827393,
657
+ "logits/rejected": 1.2642172574996948,
658
+ "logps/chosen": -27.680116653442383,
659
+ "logps/rejected": -47.270565032958984,
660
+ "loss": 1.8336,
661
+ "nll_loss": 0.1655382215976715,
662
+ "rewards/accuracies": 0.8487499952316284,
663
+ "rewards/chosen": -2.768012046813965,
664
+ "rewards/margins": 1.9590446949005127,
665
+ "rewards/rejected": -4.727056980133057,
666
+ "step": 2050
667
+ },
668
+ {
669
+ "epoch": 1.6800000000000002,
670
+ "grad_norm": 20.977998733520508,
671
+ "learning_rate": 4.8301539675328205e-06,
672
+ "logits/chosen": 1.1827168464660645,
673
+ "logits/rejected": 1.2348774671554565,
674
+ "logps/chosen": -26.652387619018555,
675
+ "logps/rejected": -46.84341812133789,
676
+ "loss": 1.758,
677
+ "nll_loss": 0.15646891295909882,
678
+ "rewards/accuracies": 0.8512499928474426,
679
+ "rewards/chosen": -2.665238857269287,
680
+ "rewards/margins": 2.0191030502319336,
681
+ "rewards/rejected": -4.684341907501221,
682
+ "step": 2100
683
+ },
684
+ {
685
+ "epoch": 1.72,
686
+ "grad_norm": 18.503433227539062,
687
+ "learning_rate": 4.597845584334387e-06,
688
+ "logits/chosen": 1.1708920001983643,
689
+ "logits/rejected": 1.2160160541534424,
690
+ "logps/chosen": -27.61713409423828,
691
+ "logps/rejected": -47.54500961303711,
692
+ "loss": 1.8257,
693
+ "nll_loss": 0.1635328233242035,
694
+ "rewards/accuracies": 0.8500000238418579,
695
+ "rewards/chosen": -2.7617132663726807,
696
+ "rewards/margins": 1.9927879571914673,
697
+ "rewards/rejected": -4.7545013427734375,
698
+ "step": 2150
699
+ },
700
+ {
701
+ "epoch": 1.76,
702
+ "grad_norm": 17.69391441345215,
703
+ "learning_rate": 4.366408178049728e-06,
704
+ "logits/chosen": 1.184067964553833,
705
+ "logits/rejected": 1.2318971157073975,
706
+ "logps/chosen": -25.899795532226562,
707
+ "logps/rejected": -45.311283111572266,
708
+ "loss": 1.7548,
709
+ "nll_loss": 0.15687091648578644,
710
+ "rewards/accuracies": 0.8550000190734863,
711
+ "rewards/chosen": -2.589979648590088,
712
+ "rewards/margins": 1.9411489963531494,
713
+ "rewards/rejected": -4.531128406524658,
714
+ "step": 2200
715
+ },
716
+ {
717
+ "epoch": 1.8,
718
+ "grad_norm": 21.093082427978516,
719
+ "learning_rate": 4.136342990565055e-06,
720
+ "logits/chosen": 1.165001630783081,
721
+ "logits/rejected": 1.2157142162322998,
722
+ "logps/chosen": -26.217182159423828,
723
+ "logps/rejected": -47.43375015258789,
724
+ "loss": 1.7791,
725
+ "nll_loss": 0.16027042269706726,
726
+ "rewards/accuracies": 0.8675000071525574,
727
+ "rewards/chosen": -2.621718406677246,
728
+ "rewards/margins": 2.121656894683838,
729
+ "rewards/rejected": -4.743375301361084,
730
+ "step": 2250
731
+ },
732
+ {
733
+ "epoch": 1.8399999999999999,
734
+ "grad_norm": 22.38056182861328,
735
+ "learning_rate": 3.908148291846225e-06,
736
+ "logits/chosen": 1.1481218338012695,
737
+ "logits/rejected": 1.1908549070358276,
738
+ "logps/chosen": -24.328386306762695,
739
+ "logps/rejected": -43.8892707824707,
740
+ "loss": 1.683,
741
+ "nll_loss": 0.1506115198135376,
742
+ "rewards/accuracies": 0.8700000047683716,
743
+ "rewards/chosen": -2.4328386783599854,
744
+ "rewards/margins": 1.9560893774032593,
745
+ "rewards/rejected": -4.388927459716797,
746
+ "step": 2300
747
+ },
748
+ {
749
+ "epoch": 1.88,
750
+ "grad_norm": 38.030303955078125,
751
+ "learning_rate": 3.6823183007969375e-06,
752
+ "logits/chosen": 1.1535931825637817,
753
+ "logits/rejected": 1.2029131650924683,
754
+ "logps/chosen": -25.072856903076172,
755
+ "logps/rejected": -44.75934982299805,
756
+ "loss": 1.7215,
757
+ "nll_loss": 0.15159070491790771,
758
+ "rewards/accuracies": 0.8424999713897705,
759
+ "rewards/chosen": -2.5072858333587646,
760
+ "rewards/margins": 1.9686492681503296,
761
+ "rewards/rejected": -4.475934982299805,
762
+ "step": 2350
763
+ },
764
+ {
765
+ "epoch": 1.92,
766
+ "grad_norm": 16.922834396362305,
767
+ "learning_rate": 3.4593421148906523e-06,
768
+ "logits/chosen": 1.150818943977356,
769
+ "logits/rejected": 1.20121431350708,
770
+ "logps/chosen": -27.409215927124023,
771
+ "logps/rejected": -46.66661071777344,
772
+ "loss": 1.7911,
773
+ "nll_loss": 0.15817226469516754,
774
+ "rewards/accuracies": 0.8450000286102295,
775
+ "rewards/chosen": -2.740921974182129,
776
+ "rewards/margins": 1.9257392883300781,
777
+ "rewards/rejected": -4.666661262512207,
778
+ "step": 2400
779
+ },
780
+ {
781
+ "epoch": 1.96,
782
+ "grad_norm": 15.894891738891602,
783
+ "learning_rate": 3.239702650894364e-06,
784
+ "logits/chosen": 1.1605645418167114,
785
+ "logits/rejected": 1.2133049964904785,
786
+ "logps/chosen": -25.806852340698242,
787
+ "logps/rejected": -46.503700256347656,
788
+ "loss": 1.739,
789
+ "nll_loss": 0.15509046614170074,
790
+ "rewards/accuracies": 0.8550000190734863,
791
+ "rewards/chosen": -2.5806851387023926,
792
+ "rewards/margins": 2.0696845054626465,
793
+ "rewards/rejected": -4.650369644165039,
794
+ "step": 2450
795
+ },
796
+ {
797
+ "epoch": 2.0,
798
+ "grad_norm": 17.814403533935547,
799
+ "learning_rate": 3.023875598978419e-06,
800
+ "logits/chosen": 1.1574615240097046,
801
+ "logits/rejected": 1.2072219848632812,
802
+ "logps/chosen": -25.269817352294922,
803
+ "logps/rejected": -47.318443298339844,
804
+ "loss": 1.6896,
805
+ "nll_loss": 0.15325431525707245,
806
+ "rewards/accuracies": 0.8812500238418579,
807
+ "rewards/chosen": -2.5269815921783447,
808
+ "rewards/margins": 2.2048630714416504,
809
+ "rewards/rejected": -4.731843948364258,
810
+ "step": 2500
811
+ },
812
+ {
813
+ "epoch": 2.04,
814
+ "grad_norm": 17.23746109008789,
815
+ "learning_rate": 2.812328392477536e-06,
816
+ "logits/chosen": 1.106400489807129,
817
+ "logits/rejected": 1.1545926332473755,
818
+ "logps/chosen": -23.484037399291992,
819
+ "logps/rejected": -48.1876106262207,
820
+ "loss": 1.499,
821
+ "nll_loss": 0.14059405028820038,
822
+ "rewards/accuracies": 0.9449999928474426,
823
+ "rewards/chosen": -2.3484036922454834,
824
+ "rewards/margins": 2.470357656478882,
825
+ "rewards/rejected": -4.818761825561523,
826
+ "step": 2550
827
+ },
828
+ {
829
+ "epoch": 2.08,
830
+ "grad_norm": 14.634954452514648,
831
+ "learning_rate": 2.6055191955342886e-06,
832
+ "logits/chosen": 1.0940419435501099,
833
+ "logits/rejected": 1.1334995031356812,
834
+ "logps/chosen": -24.663402557373047,
835
+ "logps/rejected": -49.43788146972656,
836
+ "loss": 1.5717,
837
+ "nll_loss": 0.1473802775144577,
838
+ "rewards/accuracies": 0.9275000095367432,
839
+ "rewards/chosen": -2.4663405418395996,
840
+ "rewards/margins": 2.477447509765625,
841
+ "rewards/rejected": -4.943788051605225,
842
+ "step": 2600
843
+ },
844
+ {
845
+ "epoch": 2.12,
846
+ "grad_norm": 17.942777633666992,
847
+ "learning_rate": 2.403895910817593e-06,
848
+ "logits/chosen": 1.0923943519592285,
849
+ "logits/rejected": 1.1312522888183594,
850
+ "logps/chosen": -23.995264053344727,
851
+ "logps/rejected": -47.21249008178711,
852
+ "loss": 1.5767,
853
+ "nll_loss": 0.14557117223739624,
854
+ "rewards/accuracies": 0.918749988079071,
855
+ "rewards/chosen": -2.399526357650757,
856
+ "rewards/margins": 2.3217225074768066,
857
+ "rewards/rejected": -4.721249103546143,
858
+ "step": 2650
859
+ },
860
+ {
861
+ "epoch": 2.16,
862
+ "grad_norm": 16.142436981201172,
863
+ "learning_rate": 2.2078952094652705e-06,
864
+ "logits/chosen": 1.0400832891464233,
865
+ "logits/rejected": 1.1018368005752563,
866
+ "logps/chosen": -24.179841995239258,
867
+ "logps/rejected": -50.50990295410156,
868
+ "loss": 1.5025,
869
+ "nll_loss": 0.14027239382266998,
870
+ "rewards/accuracies": 0.918749988079071,
871
+ "rewards/chosen": -2.4179842472076416,
872
+ "rewards/margins": 2.6330063343048096,
873
+ "rewards/rejected": -5.050990104675293,
874
+ "step": 2700
875
+ },
876
+ {
877
+ "epoch": 2.2,
878
+ "grad_norm": 19.593612670898438,
879
+ "learning_rate": 2.017941585351591e-06,
880
+ "logits/chosen": 1.0729751586914062,
881
+ "logits/rejected": 1.1096735000610352,
882
+ "logps/chosen": -24.066631317138672,
883
+ "logps/rejected": -47.574344635009766,
884
+ "loss": 1.5804,
885
+ "nll_loss": 0.14690618216991425,
886
+ "rewards/accuracies": 0.918749988079071,
887
+ "rewards/chosen": -2.406662940979004,
888
+ "rewards/margins": 2.3507721424102783,
889
+ "rewards/rejected": -4.757434844970703,
890
+ "step": 2750
891
+ },
892
+ {
893
+ "epoch": 2.24,
894
+ "grad_norm": 14.078289985656738,
895
+ "learning_rate": 1.8344464357280722e-06,
896
+ "logits/chosen": 1.0587913990020752,
897
+ "logits/rejected": 1.0995149612426758,
898
+ "logps/chosen": -23.153541564941406,
899
+ "logps/rejected": -47.50178527832031,
900
+ "loss": 1.4804,
901
+ "nll_loss": 0.13591693341732025,
902
+ "rewards/accuracies": 0.9375,
903
+ "rewards/chosen": -2.315354347229004,
904
+ "rewards/margins": 2.4348244667053223,
905
+ "rewards/rejected": -4.750178813934326,
906
+ "step": 2800
907
+ },
908
+ {
909
+ "epoch": 2.2800000000000002,
910
+ "grad_norm": 15.281038284301758,
911
+ "learning_rate": 1.6578071702286396e-06,
912
+ "logits/chosen": 1.0582990646362305,
913
+ "logits/rejected": 1.0928741693496704,
914
+ "logps/chosen": -23.392786026000977,
915
+ "logps/rejected": -47.97736358642578,
916
+ "loss": 1.5417,
917
+ "nll_loss": 0.1428249180316925,
918
+ "rewards/accuracies": 0.9175000190734863,
919
+ "rewards/chosen": -2.33927845954895,
920
+ "rewards/margins": 2.4584577083587646,
921
+ "rewards/rejected": -4.797736167907715,
922
+ "step": 2850
923
+ },
924
+ {
925
+ "epoch": 2.32,
926
+ "grad_norm": 15.694136619567871,
927
+ "learning_rate": 1.4884063501688539e-06,
928
+ "logits/chosen": 1.064022183418274,
929
+ "logits/rejected": 1.1182899475097656,
930
+ "logps/chosen": -23.681140899658203,
931
+ "logps/rejected": -48.91339874267578,
932
+ "loss": 1.4978,
933
+ "nll_loss": 0.1389184594154358,
934
+ "rewards/accuracies": 0.9287499785423279,
935
+ "rewards/chosen": -2.3681137561798096,
936
+ "rewards/margins": 2.523225784301758,
937
+ "rewards/rejected": -4.8913397789001465,
938
+ "step": 2900
939
+ },
940
+ {
941
+ "epoch": 2.36,
942
+ "grad_norm": 19.51872444152832,
943
+ "learning_rate": 1.3266108600032928e-06,
944
+ "logits/chosen": 1.0682127475738525,
945
+ "logits/rejected": 1.1111141443252563,
946
+ "logps/chosen": -24.015024185180664,
947
+ "logps/rejected": -47.481693267822266,
948
+ "loss": 1.5333,
949
+ "nll_loss": 0.14073985815048218,
950
+ "rewards/accuracies": 0.9125000238418579,
951
+ "rewards/chosen": -2.4015023708343506,
952
+ "rewards/margins": 2.3466668128967285,
953
+ "rewards/rejected": -4.748169422149658,
954
+ "step": 2950
955
+ },
956
+ {
957
+ "epoch": 2.4,
958
+ "grad_norm": 21.052757263183594,
959
+ "learning_rate": 1.1727711127355118e-06,
960
+ "logits/chosen": 1.054885745048523,
961
+ "logits/rejected": 1.0945353507995605,
962
+ "logps/chosen": -22.908018112182617,
963
+ "logps/rejected": -46.528038024902344,
964
+ "loss": 1.4819,
965
+ "nll_loss": 0.13584649562835693,
966
+ "rewards/accuracies": 0.9100000262260437,
967
+ "rewards/chosen": -2.290801525115967,
968
+ "rewards/margins": 2.362002372741699,
969
+ "rewards/rejected": -4.652803897857666,
970
+ "step": 3000
971
+ },
972
+ {
973
+ "epoch": 2.44,
974
+ "grad_norm": 16.752124786376953,
975
+ "learning_rate": 1.0272202910015083e-06,
976
+ "logits/chosen": 1.0628950595855713,
977
+ "logits/rejected": 1.1132755279541016,
978
+ "logps/chosen": -23.901992797851562,
979
+ "logps/rejected": -48.75297927856445,
980
+ "loss": 1.5343,
981
+ "nll_loss": 0.14184221625328064,
982
+ "rewards/accuracies": 0.9175000190734863,
983
+ "rewards/chosen": -2.390199661254883,
984
+ "rewards/margins": 2.485098361968994,
985
+ "rewards/rejected": -4.875298023223877,
986
+ "step": 3050
987
+ },
988
+ {
989
+ "epoch": 2.48,
990
+ "grad_norm": 20.135562896728516,
991
+ "learning_rate": 8.902736254703347e-07,
992
+ "logits/chosen": 1.0642120838165283,
993
+ "logits/rejected": 1.0948975086212158,
994
+ "logps/chosen": -22.418310165405273,
995
+ "logps/rejected": -44.414913177490234,
996
+ "loss": 1.4982,
997
+ "nll_loss": 0.13570523262023926,
998
+ "rewards/accuracies": 0.8999999761581421,
999
+ "rewards/chosen": -2.241831064224243,
1000
+ "rewards/margins": 2.199659824371338,
1001
+ "rewards/rejected": -4.441490650177002,
1002
+ "step": 3100
1003
+ },
1004
+ {
1005
+ "epoch": 2.52,
1006
+ "grad_norm": 17.3437557220459,
1007
+ "learning_rate": 7.622277121246513e-07,
1008
+ "logits/chosen": 1.0698988437652588,
1009
+ "logits/rejected": 1.1142799854278564,
1010
+ "logps/chosen": -23.52094078063965,
1011
+ "logps/rejected": -47.37723159790039,
1012
+ "loss": 1.5275,
1013
+ "nll_loss": 0.13999156653881073,
1014
+ "rewards/accuracies": 0.9137499928474426,
1015
+ "rewards/chosen": -2.3520936965942383,
1016
+ "rewards/margins": 2.385629415512085,
1017
+ "rewards/rejected": -4.737722873687744,
1018
+ "step": 3150
1019
+ },
1020
+ {
1021
+ "epoch": 2.56,
1022
+ "grad_norm": 19.78534698486328,
1023
+ "learning_rate": 6.433598698998766e-07,
1024
+ "logits/chosen": 1.065538763999939,
1025
+ "logits/rejected": 1.118047833442688,
1026
+ "logps/chosen": -23.858076095581055,
1027
+ "logps/rejected": -46.57390594482422,
1028
+ "loss": 1.5673,
1029
+ "nll_loss": 0.1432953178882599,
1030
+ "rewards/accuracies": 0.9024999737739563,
1031
+ "rewards/chosen": -2.385807752609253,
1032
+ "rewards/margins": 2.271583318710327,
1033
+ "rewards/rejected": -4.657390594482422,
1034
+ "step": 3200
1035
+ },
1036
+ {
1037
+ "epoch": 2.6,
1038
+ "grad_norm": 20.778789520263672,
1039
+ "learning_rate": 5.339275400731331e-07,
1040
+ "logits/chosen": 1.064239263534546,
1041
+ "logits/rejected": 1.1029696464538574,
1042
+ "logps/chosen": -23.444625854492188,
1043
+ "logps/rejected": -45.59990692138672,
1044
+ "loss": 1.5517,
1045
+ "nll_loss": 0.14206433296203613,
1046
+ "rewards/accuracies": 0.9150000214576721,
1047
+ "rewards/chosen": -2.3444626331329346,
1048
+ "rewards/margins": 2.2155280113220215,
1049
+ "rewards/rejected": -4.559990882873535,
1050
+ "step": 3250
1051
+ },
1052
+ {
1053
+ "epoch": 2.64,
1054
+ "grad_norm": 13.808253288269043,
1055
+ "learning_rate": 4.3416772870275295e-07,
1056
+ "logits/chosen": 1.0500396490097046,
1057
+ "logits/rejected": 1.0961949825286865,
1058
+ "logps/chosen": -23.120912551879883,
1059
+ "logps/rejected": -47.897361755371094,
1060
+ "loss": 1.479,
1061
+ "nll_loss": 0.13585034012794495,
1062
+ "rewards/accuracies": 0.9275000095367432,
1063
+ "rewards/chosen": -2.31209135055542,
1064
+ "rewards/margins": 2.4776451587677,
1065
+ "rewards/rejected": -4.789736270904541,
1066
+ "step": 3300
1067
+ },
1068
+ {
1069
+ "epoch": 2.68,
1070
+ "grad_norm": 13.967582702636719,
1071
+ "learning_rate": 3.442964933259474e-07,
1072
+ "logits/chosen": 1.0531511306762695,
1073
+ "logits/rejected": 1.1031429767608643,
1074
+ "logps/chosen": -23.176462173461914,
1075
+ "logps/rejected": -48.2183723449707,
1076
+ "loss": 1.4983,
1077
+ "nll_loss": 0.138520285487175,
1078
+ "rewards/accuracies": 0.9137499928474426,
1079
+ "rewards/chosen": -2.3176462650299072,
1080
+ "rewards/margins": 2.5041911602020264,
1081
+ "rewards/rejected": -4.821837425231934,
1082
+ "step": 3350
1083
+ },
1084
+ {
1085
+ "epoch": 2.7199999999999998,
1086
+ "grad_norm": 18.23953628540039,
1087
+ "learning_rate": 2.6450847502627883e-07,
1088
+ "logits/chosen": 1.038626790046692,
1089
+ "logits/rejected": 1.0861948728561401,
1090
+ "logps/chosen": -23.72449493408203,
1091
+ "logps/rejected": -49.453460693359375,
1092
+ "loss": 1.4878,
1093
+ "nll_loss": 0.13710632920265198,
1094
+ "rewards/accuracies": 0.9175000190734863,
1095
+ "rewards/chosen": -2.3724493980407715,
1096
+ "rewards/margins": 2.572896957397461,
1097
+ "rewards/rejected": -4.945346832275391,
1098
+ "step": 3400
1099
+ },
1100
+ {
1101
+ "epoch": 2.76,
1102
+ "grad_norm": 19.383655548095703,
1103
+ "learning_rate": 1.9497647688442478e-07,
1104
+ "logits/chosen": 1.0585306882858276,
1105
+ "logits/rejected": 1.1055423021316528,
1106
+ "logps/chosen": -22.849088668823242,
1107
+ "logps/rejected": -46.4496955871582,
1108
+ "loss": 1.5048,
1109
+ "nll_loss": 0.13851623237133026,
1110
+ "rewards/accuracies": 0.925000011920929,
1111
+ "rewards/chosen": -2.2849090099334717,
1112
+ "rewards/margins": 2.360060214996338,
1113
+ "rewards/rejected": -4.6449689865112305,
1114
+ "step": 3450
1115
+ },
1116
+ {
1117
+ "epoch": 2.8,
1118
+ "grad_norm": 18.49037742614746,
1119
+ "learning_rate": 1.358510897251808e-07,
1120
+ "logits/chosen": 1.0519126653671265,
1121
+ "logits/rejected": 1.0990923643112183,
1122
+ "logps/chosen": -22.153770446777344,
1123
+ "logps/rejected": -47.41790008544922,
1124
+ "loss": 1.4731,
1125
+ "nll_loss": 0.13626331090927124,
1126
+ "rewards/accuracies": 0.9212499856948853,
1127
+ "rewards/chosen": -2.21537709236145,
1128
+ "rewards/margins": 2.5264134407043457,
1129
+ "rewards/rejected": -4.741790294647217,
1130
+ "step": 3500
1131
+ },
1132
+ {
1133
+ "epoch": 2.84,
1134
+ "grad_norm": 16.536218643188477,
1135
+ "learning_rate": 8.726036597126619e-08,
1136
+ "logits/chosen": 1.0618226528167725,
1137
+ "logits/rejected": 1.1148923635482788,
1138
+ "logps/chosen": -22.412979125976562,
1139
+ "logps/rejected": -45.536277770996094,
1140
+ "loss": 1.4864,
1141
+ "nll_loss": 0.1364334225654602,
1142
+ "rewards/accuracies": 0.9162499904632568,
1143
+ "rewards/chosen": -2.241297960281372,
1144
+ "rewards/margins": 2.3123297691345215,
1145
+ "rewards/rejected": -4.5536274909973145,
1146
+ "step": 3550
1147
+ },
1148
+ {
1149
+ "epoch": 2.88,
1150
+ "grad_norm": 20.392322540283203,
1151
+ "learning_rate": 4.93095423102935e-08,
1152
+ "logits/chosen": 1.0614595413208008,
1153
+ "logits/rejected": 1.1021995544433594,
1154
+ "logps/chosen": -24.838594436645508,
1155
+ "logps/rejected": -49.431583404541016,
1156
+ "loss": 1.5749,
1157
+ "nll_loss": 0.14667731523513794,
1158
+ "rewards/accuracies": 0.9049999713897705,
1159
+ "rewards/chosen": -2.4838597774505615,
1160
+ "rewards/margins": 2.459298610687256,
1161
+ "rewards/rejected": -4.943158149719238,
1162
+ "step": 3600
1163
+ },
1164
+ {
1165
+ "epoch": 2.92,
1166
+ "grad_norm": 17.081296920776367,
1167
+ "learning_rate": 2.2080811775535006e-08,
1168
+ "logits/chosen": 1.0633502006530762,
1169
+ "logits/rejected": 1.1077544689178467,
1170
+ "logps/chosen": -24.01654052734375,
1171
+ "logps/rejected": -49.075321197509766,
1172
+ "loss": 1.5384,
1173
+ "nll_loss": 0.1433144360780716,
1174
+ "rewards/accuracies": 0.9212499856948853,
1175
+ "rewards/chosen": -2.401654005050659,
1176
+ "rewards/margins": 2.505878210067749,
1177
+ "rewards/rejected": -4.907532215118408,
1178
+ "step": 3650
1179
+ },
1180
+ {
1181
+ "epoch": 2.96,
1182
+ "grad_norm": 19.85245132446289,
1183
+ "learning_rate": 5.633145734114665e-09,
1184
+ "logits/chosen": 1.0758558511734009,
1185
+ "logits/rejected": 1.114052653312683,
1186
+ "logps/chosen": -23.172121047973633,
1187
+ "logps/rejected": -46.996238708496094,
1188
+ "loss": 1.5239,
1189
+ "nll_loss": 0.14142683148384094,
1190
+ "rewards/accuracies": 0.9262499809265137,
1191
+ "rewards/chosen": -2.3172123432159424,
1192
+ "rewards/margins": 2.3824121952056885,
1193
+ "rewards/rejected": -4.699624538421631,
1194
+ "step": 3700
1195
+ },
1196
+ {
1197
+ "epoch": 3.0,
1198
+ "grad_norm": 19.575937271118164,
1199
+ "learning_rate": 2.1661681620654963e-12,
1200
+ "logits/chosen": 1.0582020282745361,
1201
+ "logits/rejected": 1.1024978160858154,
1202
+ "logps/chosen": -22.676530838012695,
1203
+ "logps/rejected": -48.214569091796875,
1204
+ "loss": 1.4608,
1205
+ "nll_loss": 0.13524094223976135,
1206
+ "rewards/accuracies": 0.9287499785423279,
1207
+ "rewards/chosen": -2.267652988433838,
1208
+ "rewards/margins": 2.5538036823272705,
1209
+ "rewards/rejected": -4.821456432342529,
1210
+ "step": 3750
1211
+ }
1212
+ ],
1213
+ "logging_steps": 50,
1214
+ "max_steps": 3750,
1215
+ "num_input_tokens_seen": 0,
1216
+ "num_train_epochs": 3,
1217
+ "save_steps": 200,
1218
+ "stateful_callbacks": {
1219
+ "TrainerControl": {
1220
+ "args": {
1221
+ "should_epoch_stop": false,
1222
+ "should_evaluate": false,
1223
+ "should_log": false,
1224
+ "should_save": true,
1225
+ "should_training_stop": true
1226
+ },
1227
+ "attributes": {}
1228
+ }
1229
+ },
1230
+ "total_flos": 0.0,
1231
+ "train_batch_size": 2,
1232
+ "trial_name": null,
1233
+ "trial_params": null
1234
+ }
training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9aa02136503d05933309aa9f771cfd1d85312a3212ad3e32671f02280841e7f4
3
+ size 6225
vocab.json ADDED
The diff for this file is too large to render. See raw diff