cimol commited on
Commit
cf5adaf
·
verified ·
1 Parent(s): 7e50812

Training in progress, step 300, checkpoint

Browse files
last-checkpoint/adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:e90b987879ddf480be0ca4703eefea597215d15c71ab1ef1fb817ce4ed32ec57
3
  size 335604696
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:07d192aa1c9ce20fd264be177399627c549841ba4a3dcbe785cfdc6ff6e4acf4
3
  size 335604696
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:c2af4a2686e254f4f87cd7bd27743c9d282b16715e9d9a2349888409f80d297a
3
- size 170920084
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3a08f49a47f3ea1d7622dfb86c37c67c48b24de609d9cf949a91fafec68666c5
3
+ size 170920532
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:8a355a58987eda105cf552ad712e56925deafbd21c52f01db6bb79781434e2f1
3
  size 14244
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:616d5591fe2815fb845a002baa0bbc69990e73ea586e4a98cd405180dd111c49
3
  size 14244
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:7e080d0e7c071615c66cdfbd19fb773bb51f3a695898c5afa82d7150c8b34389
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:35792e2ed51bf6bc8a2a5bf1bcfa90fe1bca1ed12b3ff3e83b15d68253c4d18f
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
- "best_metric": 0.9399939775466919,
3
- "best_model_checkpoint": "miner_id_24/checkpoint-150",
4
- "epoch": 1.2765957446808511,
5
  "eval_steps": 50,
6
- "global_step": 150,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
@@ -144,6 +144,135 @@
144
  "eval_samples_per_second": 13.215,
145
  "eval_steps_per_second": 3.337,
146
  "step": 150
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
147
  }
148
  ],
149
  "logging_steps": 10,
@@ -172,7 +301,7 @@
172
  "attributes": {}
173
  }
174
  },
175
- "total_flos": 2.2543503351349248e+17,
176
  "train_batch_size": 8,
177
  "trial_name": null,
178
  "trial_params": null
 
1
  {
2
+ "best_metric": 0.7150752544403076,
3
+ "best_model_checkpoint": "miner_id_24/checkpoint-300",
4
+ "epoch": 2.5531914893617023,
5
  "eval_steps": 50,
6
+ "global_step": 300,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
 
144
  "eval_samples_per_second": 13.215,
145
  "eval_steps_per_second": 3.337,
146
  "step": 150
147
+ },
148
+ {
149
+ "epoch": 1.3617021276595744,
150
+ "grad_norm": 1.4046274423599243,
151
+ "learning_rate": 0.00011956517382403321,
152
+ "loss": 0.7635,
153
+ "step": 160
154
+ },
155
+ {
156
+ "epoch": 1.4468085106382977,
157
+ "grad_norm": 1.1550798416137695,
158
+ "learning_rate": 0.00011051357095705101,
159
+ "loss": 0.6612,
160
+ "step": 170
161
+ },
162
+ {
163
+ "epoch": 1.5319148936170213,
164
+ "grad_norm": 1.0277212858200073,
165
+ "learning_rate": 0.00010137383111821266,
166
+ "loss": 0.9006,
167
+ "step": 180
168
+ },
169
+ {
170
+ "epoch": 1.6170212765957448,
171
+ "grad_norm": 0.9155287742614746,
172
+ "learning_rate": 9.22225742301153e-05,
173
+ "loss": 0.7741,
174
+ "step": 190
175
+ },
176
+ {
177
+ "epoch": 1.702127659574468,
178
+ "grad_norm": 1.3166751861572266,
179
+ "learning_rate": 8.313651676464923e-05,
180
+ "loss": 0.6244,
181
+ "step": 200
182
+ },
183
+ {
184
+ "epoch": 1.702127659574468,
185
+ "eval_loss": 0.7949697375297546,
186
+ "eval_runtime": 14.9637,
187
+ "eval_samples_per_second": 13.232,
188
+ "eval_steps_per_second": 3.341,
189
+ "step": 200
190
+ },
191
+ {
192
+ "epoch": 1.7872340425531914,
193
+ "grad_norm": 0.9204571843147278,
194
+ "learning_rate": 7.419182861636218e-05,
195
+ "loss": 0.8257,
196
+ "step": 210
197
+ },
198
+ {
199
+ "epoch": 1.872340425531915,
200
+ "grad_norm": 1.011165976524353,
201
+ "learning_rate": 6.546349455786926e-05,
202
+ "loss": 0.6789,
203
+ "step": 220
204
+ },
205
+ {
206
+ "epoch": 1.9574468085106385,
207
+ "grad_norm": 1.240622878074646,
208
+ "learning_rate": 5.702468563033306e-05,
209
+ "loss": 0.5821,
210
+ "step": 230
211
+ },
212
+ {
213
+ "epoch": 2.0425531914893615,
214
+ "grad_norm": 0.803183913230896,
215
+ "learning_rate": 4.894614573874877e-05,
216
+ "loss": 0.8176,
217
+ "step": 240
218
+ },
219
+ {
220
+ "epoch": 2.127659574468085,
221
+ "grad_norm": 0.8154023289680481,
222
+ "learning_rate": 4.129559859430573e-05,
223
+ "loss": 0.4854,
224
+ "step": 250
225
+ },
226
+ {
227
+ "epoch": 2.127659574468085,
228
+ "eval_loss": 0.7357346415519714,
229
+ "eval_runtime": 14.9479,
230
+ "eval_samples_per_second": 13.246,
231
+ "eval_steps_per_second": 3.345,
232
+ "step": 250
233
+ },
234
+ {
235
+ "epoch": 2.2127659574468086,
236
+ "grad_norm": 1.3029338121414185,
237
+ "learning_rate": 3.413717997552376e-05,
238
+ "loss": 0.3565,
239
+ "step": 260
240
+ },
241
+ {
242
+ "epoch": 2.297872340425532,
243
+ "grad_norm": 1.1356505155563354,
244
+ "learning_rate": 2.7530900067611577e-05,
245
+ "loss": 0.6265,
246
+ "step": 270
247
+ },
248
+ {
249
+ "epoch": 2.382978723404255,
250
+ "grad_norm": 1.0218287706375122,
251
+ "learning_rate": 2.1532140387343735e-05,
252
+ "loss": 0.43,
253
+ "step": 280
254
+ },
255
+ {
256
+ "epoch": 2.4680851063829787,
257
+ "grad_norm": 0.9909137487411499,
258
+ "learning_rate": 1.619118951081594e-05,
259
+ "loss": 0.3348,
260
+ "step": 290
261
+ },
262
+ {
263
+ "epoch": 2.5531914893617023,
264
+ "grad_norm": 1.1617368459701538,
265
+ "learning_rate": 1.1552821496149135e-05,
266
+ "loss": 0.6773,
267
+ "step": 300
268
+ },
269
+ {
270
+ "epoch": 2.5531914893617023,
271
+ "eval_loss": 0.7150752544403076,
272
+ "eval_runtime": 14.938,
273
+ "eval_samples_per_second": 13.255,
274
+ "eval_steps_per_second": 3.347,
275
+ "step": 300
276
  }
277
  ],
278
  "logging_steps": 10,
 
301
  "attributes": {}
302
  }
303
  },
304
+ "total_flos": 4.5087006702698496e+17,
305
  "train_batch_size": 8,
306
  "trial_name": null,
307
  "trial_params": null