nttx commited on
Commit
2b0ec09
·
verified ·
1 Parent(s): 0f9688e

Training in progress, step 3000, checkpoint

Browse files
last-checkpoint/adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:4024669463897c198606d97e6d594db0d110cc0386bf2567c16a1d0c4bdbdd37
3
  size 60576160
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:310cb3c19ba8648eac2cfc5efd2edfa9eb7aef6e777d68c2791adcdb29bcfe75
3
  size 60576160
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:a6fd26d675fa6a6a126306338b9f4505353715b9e6c6a2cf270cbd3fef711479
3
  size 30896058
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c0e117a815755828fbf50e88be104a33e767c27b104de7a015ba811e06e6db5b
3
  size 30896058
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:d5fb513b56a15fd673084e1d38e8da5300b1258dad221652d1257fc6c613c04a
3
  size 14244
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f6650480ac19d065a63aa287d558a6455d718894f4c96375f8466f0bd81805e5
3
  size 14244
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:5f76a102ac755d5bd19e6f0e55ad50e49a03a307409b32eeac7fb5a2a13752ba
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4b5b8ae682ae9c51a247d395336eab40b1fa2d190384873d8421e390f09af11a
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
- "best_metric": 4.839520454406738,
3
- "best_model_checkpoint": "miner_id_24/checkpoint-2850",
4
- "epoch": 0.934043424825891,
5
  "eval_steps": 150,
6
- "global_step": 2850,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
@@ -566,6 +566,35 @@
566
  "eval_samples_per_second": 247.137,
567
  "eval_steps_per_second": 61.832,
568
  "step": 2850
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
569
  }
570
  ],
571
  "logging_steps": 50,
@@ -589,12 +618,12 @@
589
  "should_evaluate": false,
590
  "should_log": false,
591
  "should_save": true,
592
- "should_training_stop": false
593
  },
594
  "attributes": {}
595
  }
596
  },
597
- "total_flos": 3.95906974262231e+16,
598
  "train_batch_size": 4,
599
  "trial_name": null,
600
  "trial_params": null
 
1
  {
2
+ "best_metric": 4.8373589515686035,
3
+ "best_model_checkpoint": "miner_id_24/checkpoint-3000",
4
+ "epoch": 0.9832036050798852,
5
  "eval_steps": 150,
6
+ "global_step": 3000,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
 
566
  "eval_samples_per_second": 247.137,
567
  "eval_steps_per_second": 61.832,
568
  "step": 2850
569
+ },
570
+ {
571
+ "epoch": 0.9504301515772224,
572
+ "grad_norm": 8.805246353149414,
573
+ "learning_rate": 5.665199789862907e-07,
574
+ "loss": 4.9501,
575
+ "step": 2900
576
+ },
577
+ {
578
+ "epoch": 0.9668168783285539,
579
+ "grad_norm": 10.70975112915039,
580
+ "learning_rate": 1.4173043232380557e-07,
581
+ "loss": 4.9379,
582
+ "step": 2950
583
+ },
584
+ {
585
+ "epoch": 0.9832036050798852,
586
+ "grad_norm": 8.567855834960938,
587
+ "learning_rate": 0.0,
588
+ "loss": 4.9035,
589
+ "step": 3000
590
+ },
591
+ {
592
+ "epoch": 0.9832036050798852,
593
+ "eval_loss": 4.8373589515686035,
594
+ "eval_runtime": 10.38,
595
+ "eval_samples_per_second": 247.592,
596
+ "eval_steps_per_second": 61.946,
597
+ "step": 3000
598
  }
599
  ],
600
  "logging_steps": 50,
 
618
  "should_evaluate": false,
619
  "should_log": false,
620
  "should_save": true,
621
+ "should_training_stop": true
622
  },
623
  "attributes": {}
624
  }
625
  },
626
+ "total_flos": 4.168168593004954e+16,
627
  "train_batch_size": 4,
628
  "trial_name": null,
629
  "trial_params": null