nttx commited on
Commit
152befd
·
verified ·
1 Parent(s): ecab09a

Training in progress, step 1650, checkpoint

Browse files
last-checkpoint/adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:f2436728f82aa4949abbbd0eb1a6a832a2bcc84241a1c8d50f4a85aee88f4ed2
3
  size 39131224
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:12dd6337fff44ebfc4e7e1f6faec086a74126aed257725a219090db632ba1030
3
  size 39131224
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:7b97c4097242b845374e8cda7310f7388079dacabc54c6fd4bd420325b61d051
3
  size 78510334
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4678aae90eb9fa309f5a4a27ba89b7a9b22d8a4eec881787c8813398426c1d77
3
  size 78510334
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:74c141c60220302a6e2d4f05b7f65149392bf2b0c2708ca84f4fa67ea3c279a4
3
  size 14244
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:220f636333a5625ead2a747eab34ed7373b77b6069afb40aaf422414ec699f94
3
  size 14244
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:befb5de8641401f08357d93e09a6b263a413b6c62f0e4dc381df697735030b77
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:090e504a81765b504142cff403afcf92b817f0e48945f0e4e014e79a6565dad0
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
- "best_metric": 0.7997989654541016,
3
- "best_model_checkpoint": "miner_id_24/checkpoint-1500",
4
- "epoch": 2.2263030977555185,
5
  "eval_steps": 150,
6
- "global_step": 1500,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
@@ -305,6 +305,35 @@
305
  "eval_samples_per_second": 40.629,
306
  "eval_steps_per_second": 20.315,
307
  "step": 1500
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
308
  }
309
  ],
310
  "logging_steps": 50,
@@ -333,7 +362,7 @@
333
  "attributes": {}
334
  }
335
  },
336
- "total_flos": 1.7100828573696e+16,
337
  "train_batch_size": 2,
338
  "trial_name": null,
339
  "trial_params": null
 
1
  {
2
+ "best_metric": 0.7954422831535339,
3
+ "best_model_checkpoint": "miner_id_24/checkpoint-1650",
4
+ "epoch": 2.4488963086625857,
5
  "eval_steps": 150,
6
+ "global_step": 1650,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
 
305
  "eval_samples_per_second": 40.629,
306
  "eval_steps_per_second": 20.315,
307
  "step": 1500
308
+ },
309
+ {
310
+ "epoch": 2.300500834724541,
311
+ "grad_norm": 0.18251581490039825,
312
+ "learning_rate": 9.733794785622253e-05,
313
+ "loss": 0.7668,
314
+ "step": 1550
315
+ },
316
+ {
317
+ "epoch": 2.374698571693563,
318
+ "grad_norm": 0.1793128401041031,
319
+ "learning_rate": 9.202138944469168e-05,
320
+ "loss": 0.7456,
321
+ "step": 1600
322
+ },
323
+ {
324
+ "epoch": 2.4488963086625857,
325
+ "grad_norm": 0.17612189054489136,
326
+ "learning_rate": 8.672744727162781e-05,
327
+ "loss": 0.7508,
328
+ "step": 1650
329
+ },
330
+ {
331
+ "epoch": 2.4488963086625857,
332
+ "eval_loss": 0.7954422831535339,
333
+ "eval_runtime": 14.3584,
334
+ "eval_samples_per_second": 39.559,
335
+ "eval_steps_per_second": 19.779,
336
+ "step": 1650
337
  }
338
  ],
339
  "logging_steps": 50,
 
362
  "attributes": {}
363
  }
364
  },
365
+ "total_flos": 1.88109114310656e+16,
366
  "train_batch_size": 2,
367
  "trial_name": null,
368
  "trial_params": null